var/home/core/zuul-output/0000755000175000017500000000000015153607651014536 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015153626513015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000336006115153626365020273 0ustar corecore,ikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ Eڤ펯_ˎ6_o#oVݏKf핷ox[o8W5-b6"οƼ>UWm׫Y_?|uݗ[y[L-V_pY_P-bXwûxwAۋt[~ _P^~&RY,yDy~z]fs,l<L& " d :o5J=nJw1f /%\xiƙQʀClxv< |N ?%5$) y5? fۮ?tT)x[@Y[`VQYY0gr.W9{r&r%LӶ`zV=Too|@E1%]˜(O)X(6I;Ff"mcI۫d@FNsdxό?2$&tg*Y%\ߘfDP'F%Ab*d@e˛H,љ:72 2ƴ40tr>PYD'vt'oI¢w}o٬owko%gQ(%t#NL֜ eh&Ƨ,RH 4*,!SD 1Ed_wkxdL3F;/u7Taqu5Ոӄp\2dd$YLYG(#?%U?hB\;ErE& SOZXHBWy|iZ~hal\t2Hgb*t--ߗ|Hp(-J C?>:zR{܃ lM6_Oފ?O1nԝG?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;_x+Vy<h\dN9:bġ7 -Pwȹl;M@n̞Qj_P\ Q]GcPN;e7Vtś98m1<:|a+.:a4nՒ,]LF0);I$>ga5"f[B[fhT/ɾg}\Sj#3hEEH*Nf äE@O0~y[쾋t=iYhșC 5ܩa!ǛfGtzz*з 55E9Fa?Zk80ݞN|:AОNo;Ⱦzu\0Ac/T%;m ~S`#u.Џ1qNp&gK60nqtƅ": C@!P q]G0,d%1}Uhs;H?)M"뛲@.Cs*H _0:P.BvJ>mIyVVTF% tFL-*$tZm2AČAE9ϯ~ihFf&6,֗&̴+s~x?53!}Z[F)RH?uvͪ _5l *7h?cF_]CNnW)F5d,0SSNK9ް4:ÒozsB<^+鄌4:B%cXhK I}!5 YM%o<>"ہ)Za@Ι}YJz{ɛr|hxY/O$Zøu32EʉD'MS1}t i:Y`cФIX0$lη˽`!i:ګPSPٔ3@5;ȕ}PkڪH9' |":", 1Ҫ8 %lg&:2JC!Mjܽ#`PJWP4Q2:IGӸۡshN+60#:mufe߿~Y,iǑ wVq*T+ w%fx6 %u̩1hӰc%AYW ZY~a_6_yWf`rVA,f=A}h&VOK(BSsǽҰ%>kh5nIYk'LVc(a<1mCޢmp.֣?5t罦X[nMcow&|||x:k/.EoV%#?%W۱`3fs䓯ҴgqmubIfp$HhtLzܝ6rq/nLN?2Ǒ|;C@,UѩJ:|n^/GSZ;m#Nvd?PqTcLQMhg:F[bTm!V`AqPaPheUJ& z?NwpGj{VjQS,؃I'[y~EQ(S +mpN, Mq 70eP/d bP6k:Rǜ%V1Ȁ Z(Q:IZaP,MI6o ޞ22ݡjR:g?m@ڤB^dh NS߿c9e#C _-XѪ;Ʃ2tStΆ,~Lp`-;uIBqBVlU_~F_+ERz#{)@o\!@q['&&$"THl#d0 %L+`8zOҚƞ`wF~;~pkѽ)'cL@i]<ք6ym®Yi&s`dyMX](^!#h k:U7Uv7чd)wB5v-)s蓍\>S[l52, 5 CۈP$0Zg=+DJ%D  *NpJ֊iTv)vtT̅Rhɇ ќuގ¢6}#LpFD58LQ LvqZDOF_[2ahwfm#Y~!%rpWMEWMjbn(ek~iQ)à/2,?O 0D"\KjPQ>Y{Ÿ>14`SČ.HPdp12 (7 _:+$ߗv{wzM$VbήdsOw<}#b[E7imH'Y`;5{$ь'gISzp; AQvDIyHc<槔w w?38v?Lsb s "NDr3\{J KP/ߢ/emPW֦?>Y5p&nr0:9%Ws$Wc0FS=>Qp:!DE5^9-0 R2ڲ]ew۵jI\'iħ1 {\FPG"$$ {+!˨?EP' =@~edF \r!٤ã_e=P1W3c +A)9V ]rVmeK\4? 8'*MTox6[qn2XwK\^-ޖA2U]E_Dm5^"d*MQǜq؈f+C/tfRxeKboc5Iv{K TV}uuyk s" &ﱏҞO/ont~]5\ʅSHwӍq6Ung'!! e#@\YV,4&`-6 E=߶EYE=P?~݆]Ōvton5 lvǫV*k*5]^RFlj]R#Uz |wmTeM kuu8@8/X[1fiMiT+9[ŗ6 BN=rR60#tE#u2k *+e7[YU6Msj$wբh+8kMZY9X\u7Kp:׽ ^҃5M>!6~ö9M( Pnuݮ)`Q6eMӁKzFZf;5IW1i[xU 0FPM]gl}>6sUDO5f p6mD[%ZZvm̓'!n&.TU n$%rIwP(fwnv :Nb=X~ax`;Vw}wvRS1q!z989ep 5w%ZU.]5`s=r&v2FaUM 6/"IiBSpp3n_9>Byݝ0_5bZ8ւ 6{Sf觋-V=Oߖm!6jm3Kx6BDhvzZn8hSlz z6^Q1* _> 8A@>!a:dC<mWu[7-D[9)/*˸PP!j-7BtK|VXnT&eZc~=31mס̈'K^r,W˲vtv|,SԽ[qɑ)6&vד4G&%JLi[? 1A ۥ͟յt9 ",@9 P==s 0py(nWDwpɡ`i?E1Q!:5*6@q\\YWTk sspww0SZ2, uvao=\Sl Uݚu@$Pup՗з҃TXskwqRtYڢLhw KO5C\-&-qQ4Mv8pS俺kCߤ`ZnTV*P,rq<-mOK[[ߢm۽ȑt^, tJbظ&Pg%㢒\QS܁vn` *3UP0Sp8:>m(Zx ,c|!0=0{ P*27ެT|A_mnZ7sDbyT'77J6:ѩ> EKud^5+mn(fnc.^xt4gD638L"!}LpInTeD_1ZrbkI%8zPU:LNTPlI&N:o&2BVb+uxZ`v?7"I8hp A&?a(8E-DHa%LMg2:-ŷX(ǒ>,ݵ𴛾é5Zٵ]z"]òƓVgzEY9[Nj_vZ :jJ2^b_ F w#X6Sho禮<u8.H#',c@V8 iRX &4ڻ8zݽ.7jhvQ:H0Np: qfՋ40oW&&ף \9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2oԞ!оG/~oW(91ݧ$uxp/Cq6Un9%Z`.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og\IGֻq<-uˮ◶>waPcPw3``m- } vS¢=j=1 W=&;JW(7b ?Q.|K,ϩ3g)D͵Q5PBj(h<[rqTɈjM-y͢FY~p_~O5-֠kDNTͷItI1mk"@$AǏ}%S5<`d+0o,AրcbvJ2O`gA2Ȏp@Z#"U4Xk1G;7#m eji'ĒGIqB//(O &1I;svHd=mJW~ړUCOīpAiB^MP=MQ`ʻfRBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ63" ۩:6=TZõ$E,ϓRV|G&$rr;J TtIHFE=RȬ]P pLm|?$%>Eü%mWO[>Xmw,*9.[G n >X8Ī;xW%dT:`ٓ~:QO,}j6j!yڦʲT:Pqҋh] H+&=>g| Z;D8ܶb:! Å{2:+au 6:!fF+0#+̬NY"!6a7#񕪰%:r|o5Znڧs?si/W qEU馥˟^_޶oڷOj'?nc]Rn\t3^邳塨Lɏ"8k8M~?M}OAH$77f|lgn I;.K*!<+"eK5c&`X:#;@B@[(K44sBFu M.MNWLlY]K᜴=/ VމYlϿ4i36$>m|_>9|dUA"{!$jKx E$K3hN(tÊ-#v#O N, 9g80Ǭ&VdӞ5W1!1KYd`,-*&>F~⯰&jb.~cNk BL_OG]Bv.A|'qT(Ol.' 4IE|@Iі)<-p JkQm1 `qacܗVc?)cl*&<}P媠E{-sVU>߇GUt\+n3X]Byoz)li$2cPs6D>TE-n# rve{椱I |p)U݋7yJw&PzDgi xs  xh\L r Ѥo Zt(I >|$>tnMdэo`cZV yBMHVuV_K2k*`cKxuBG&24T}Lai 0Va(7K#ӊ!,ZDxFQO*lם>!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜˻ x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d->'I`{W6Ҥ/%ؼ:N&cLxw2Fݔ5~HJϖULXI:\D 8;J"<<_DUa2jSW%z<- H`_"4xFEW}S RD@%d䋽OHaoeДU>onueg<*_jDUaj/B}V T콵`_b5Y@V3ľuX.H fʃE-"u&L . ]Ub#}Z`.>9J}@<8/2,o;%e#FwKF؟a~dri|1g!cKgy|- д񲸲t_WHgzʴ",B=2XGSpڡ) 5鸑"oiͰm`\n= +Z?rAzB0$_A3Wcrt ~j=j䈢I&0<}wÓDֿ;<" ҽn;vBKfz [IÔ/H'o"a^~C_ħd-D_>'ӅХZCYL0=_<!ط( :"|wOS^L斦kC%&~SȐ;*-|2a_ƅa]΅_ԙgv_yT4[aYﵠ*=(oD9`7*p u\%Bɞں;}zqL7S7\۵_"'_*tW^{0Kc\bC|Y~V@;;K{B^5_uNn{d,gF}Ea°NXLl@%"Uun*/w-3Pl; }f5</Z+ʼ(U?3 ,uxyiK<VGRXaf1mKҍ<߈B{^3?&y4o<3 ® >LrwGY&[SIdY 2J^`<2O ol= B^!X~ -o 4mŚVƩd&<E)0k)^3QKbj7 OP֨*d_Gt꨷o,ѹY]yXMj^]Vv`,qKk]rsޥ?ZxfhjZgU v_r w Jqʧa\=%`NOyw݇ Ŋg"ȯG5%<||FD |ylOts}ctGg##i`SJ:NQ^F# ѳ4Q^'U3zYB5\u;3w䨍Sq4FMyzt_Kq|gu{FIP>O>Mgc^?&G岭AoyÛl :^#œץINs!I`|k{ onq׷ןӇSu=Π`K$9ysz 2ܭL{G4–9Ft6Ы*UAuTUYa{q*O M=Vb)q h5p] ~ ;lEzceأ0-G$*CI+\O 1Sr>z\[6>24lXֆNoN(>޽uW60vP d`O6ywM`7sg.0+9 C`;ϐwOѽ (a  @;yM fp0wg@q֕z~I;.ݿ n[߅3v  A/,PoVƲvd@d[; n>'r[s1-y~XM0g~ĭغ fZ#9EM9$YVXB+oԐ@;r=@qްP5J ۍBl7SiG7_/!e>M#a`^3pc%o|35?7~P+Q !mo+M)7]JAy[[E?w?a^6f7E]! sw5mJ_!ýڑHJ"4mwmڼwdINX> )ٴxSiԻnpf$[ {Ϋ j7%(XXZ==pళ+ ]$T70 0>5Z{?DG9۞LLu~'[P7B)_f;ˉ+|,`xax8hؑT ~/U?db[_:":;BaKg>F7 y)'Pg?O+w&U*Nl8SETgN1IV?SDsX=\կt\A# ?)NuOj Wٿ~5%`.{^e 'af:=ihGqU7' c/L8:;hډ!_<-ϼMt+0M:?J+C~ᾃ?>wu2J[6=+€0Lec[O蝱]IPב{`";wم+anq(yq0׿NC pj0֡g@V݄JPB>DvX>\3sJˡ8{];`0ѵ^`)ok@ĀmM8,܎vq= P$D e/6 0<!NzP7½$0@0#f"֪S@pn)"@~2 _tdv=q"BSa(ڕ9Xi;nծ `yEp3EK(J/i+ /x`qZ+N]"hqLދJv# X(m0Pq(Y-R2Z#эabȷy p&߿\z>nA픑@PI ǩ{OY4m ݎY$ѕ ,{O(E\&,)*lg6'lu-8u_8 K* ΂6 hgrW>@-B%ZJ" ,N[ EtD[·h7ZDuډAKD +,:AЉ5ln؀DgNE E'Dv%Xvv/KeWNHZ~nWY(AG2Z8q?[Z\.]vpt]$i?I\N۔a ޒMΖuӎ8 Jbm>Il=I/ 6Y{Zٓ6zֳ4Z|:ݼE|I-#u,)LHHC$u(> $&s8aHY-+R_\QL) %HqܡoWpۼ}C 9,!YY;bHWԉJ]iH(XDD20#B]\ڣ~.C?:dT\^faL}<]=<RΗaEjnN$" @(5)r` mIs;hBI ha{_{2 8f5o~*g];kpۘ2V-dZ3%XunL^]kZ}򂉏͕ V}?/+z$p jJQ6x#5 c=Q ِS]$1w_$ơ!8AdݒzXg elif^BQM9rj133}#7xKt~i0 o=#Q1I{?, (pZZJ(9D&$)O,eFe8V Q?;W養X',OuQ9՞9_[zۯ>;xDɄ0A_2hY2+ q"* xnzȢCxl(;`kٝ$!B,TZKLH,Mt`zhO)Us+ikSñ:A "8TzcUT 悔7 8%4{HYOaڣAS!A!ɳ3H͛ - a>'gW#X<AC7D"P䗸1Xt|C7|M2WlA/“|30{Bdet%|*y ʈIy gDDEg,؊_r_J O7 !>sb2%74 Y=ߖ Z%(Z`GđD_e\[[;6][' +o wJ8II٣T܌{p]OpT">:VMիu=> 2/& TL 8Il;JҾn躓OWpLO^6xj3IW1nQAkY9ӧuE}ʩ˞V(;'m af]EۤW)cᷦ92zFDBm]Ro10ގ=uXuu`٧ cft2t bHOnA([&mA(۞P쉄-˄-w#?P2FDB- 4؞`7B'lAhLh>p B2b BO$TlA\&TnAܞP5.R)ke _H-c| XChmLyˆMyΓ2pa&/FFnU1=xJiѩ}HuʬH\f/+}ך!_;FۯxO6j=Nþi? ޿E{X5Oax-0)*#Ne>pXK={.yӠc͟ƚ±_=5rVb6GR{\qjM"B/K}yx)y?q*KIؠI3E(LE`j 5C KdS/JP.Ust b$؄9+C3:)Β)PECФ<;8WNPTկ_$K`xvi\֖ -x| #R@ 㙣eCd?շw+>hHlh:fͣ"=}~d sߝʪWYY8sfiݹ1anpM挊"AdIcgٰ2RuB&*nfCz&/foprdOmj=riGj;l˦r~Cev:;**Y8oMc!ڷQ:ׂ!1GzU'uYWv!= td7L5KxC,v3w w^a ϼ*.*އ3e>M4Gzm!l蒞q~gw7**]HSfÑEڝBjCQMG7T#ﵫA2dzOD~Φd+^UOp g ;H ǧ>@ݿPI]eՌG?Geowshw#?Oοùm~]ǹ=qzƕU>PڣQxGMsa8i>g nH񵺅Q gՄԱ۟JQ4Dr)+kTq<ƢC$^ee$jD[CKQ ;HEcgBjH"hBT!_sJa!WeH]9NfP$ګo̸J_6A^D}1f,=r'& fo+iLu>Aa lp.z>)6ܫߓja0++꯵f̌ s* 1 TtM "59[wf.zhn)޼w7&+] l0$Qdnke.BcrfDsOduьO}x'ѩ Y(0Vjd' 0Ʌ(pp0O'^Ze%F,dD=A,8 /!"zvC_K.kQh˦6B+)-A=>-6|bbƗ|pJ7d}_:z*YluyM[W?uV3U/[?gf˚ }OE/)x 0}WZ{ՙ Wjä(J'Ֆc0rgLweڜƷmLn(Ϧ'Ah +vuݒhw B\-sk񽖴$NMɻ޶W :䵘G"-y4KR#.<5V/. $y]uj]#ܟ)[v:4D&_gj$~CTF܄-pX_C9l>'̸E~_~cm\}K%Woo?{YՏޟ_M)no>]5nxU՟')&PvLJLJnꪳO/2Fm }߳l3 ~|WCWstɟ]Ln>'vtD;h~cz}ۿ&evȜ㝭mޚ'd~:[-p?9gOI.dFZO><ȻmgKL{ K8NN/>LʕإkdtŚ31%"O¥,I@^C͹^ m&iW8>|,`Z _d$jMG!s#M ljp:d &KmUX[ ^Þ5l==uB'> fpV 3˴Պ\ T gI'G  wq(m{ɟC? M SdTTpc3 I'FRm Y|1VE^jK@"F -e5q Hդ3&3wJz^ # 8s𵇤,G1BHhMh] ^Wcleʇ6: {OQia]ȍ0sU$!4nY(`i4wb$Fz,hcdv,g 4.y9S\!"qH:1m1X8B'5bhucbZb w*b$Yaן8 Aj.[M)Mp!cICtW $ Zkd#F҉>]DrQ!Dzq}&} 4k*!S*8%= `^+MJ:?Q`*djT|MɁMٻ.Vblta $"1mf mAac,WZS0 J.#È! >hNuAn{ ulQ5T5cڬZc9H%.`Z)Fh+#i@ Fm\/kawY.U:2:&i p\D8ptE%b4A/EscՅ/Myttn)4WcӡfȿX/D׉dE֣1NyKtD"|zi le ksB[ Φ1K#?Ú!.F%ɦDVo^(uzC׸Rd":/XSN{*5jq&u? f\63Og2| H_?RTGyQ"qk^}ab{éI78޵dS(kBc$V(n_$87~vZ:1m֛g2sO΍ttS~ms-B?Ty 04>p2w\lTh\! m[١3C iGL ɐ<Ax ;0Nοdz%QE1>aL3Qsw$c _藕$0II=im2.0UbZ 盉V#tH~k2^,k(yvJ&c^.FǛ5 N]) xȒ¸y`փb5K6&mQE0NMYyzƸ3=5k؁J:x&Jp\ER>KTCΥ`$2}DЕg2gTʃh$N%JZ\aߖD `|nsD/`*ط fXA 0%r #8 2Pd ŽR&X 98L?bt6iHUBǥ]]nϏ$8&+A: as(dnlh}VX+xI88$)n7zjW N F/fCK,ИHĪMD}bXuՊC3m`>8kF( GΎa$w$j({:`m8%8@a}aVW>ORp=|Lw f;twK7feN}w)dt,'VؙL%r1X_Dyfq~xkfI`ܓz.s,0^sÿ!+Nn_ʊ7/.i>?g #@Mպ폽Cf!*mDH:f(/wG9xGdnKԯg!i!%-111۾||^r<%ވg\G$h{s[f$Hd˵rQal8)>,ΑY1vr *)*L+X ٳVRZf[ 抑tq#ËE5.<1C rm9ȥI/[GL~ל41G-Wyvqd(ޘW0MSfp\_<`|I'q q pZ 2$\ xKjɑWϓe%7&?kIaП@>>ǧU^d G~ill"SȠ^shh.SFolY>^R\ww֧SH9xO/>uqĭ#f,qK#͆d +$#KO𮻵:=:%K,^+ ZpTi%@v?=w_ 6uKXzsiu{#_HTAyMWUd4+)&Ҋ,΄X WjeMcdnəB,5E`LRtcjwa`7x\Y3n\I98j,Y 雂EU[؄t׀j0AHTjYF|I0u(clc54u.8].S|`qp?ìGQìԧ^/*8 N%wbbB3+LezeJwkFFcYnoƨrf<#д 9dU@3ä)k\E0IzuuR= `s7`=0/O.8'2Ykb:Db\$º!jyF6:z٧#$X6s%fQs@8V *d lpPAw tRz*$8j(v1cVw}=;lRN]rNwز)ve3EaZǴոI {0g!hLR M69.yH:%+ Oۯ!(( (Ȫ:BĖ0]L:;pZ^֫L3ڤFnf)ôwy. !tcUK$a1F,AJl^3A7ީhp6RdtpǾT![̈mJ[mb@_-ErbYGcF:U ͞΀:C4E9'/ׇP*E7{#Q:/úw 3c@H4rQ J3 u27*iu"?6ǀi!XUa0uw_zH򒣘TG*e5 nH$؍.m:'oԣ~ϽkBHs?fHŗIpwя{iŢCfbC(i-xD {ȑJ$u (Nv2Oͺ홐UyඹئmvOї*G_ð1N䂽=9\nGWTmn{1KI6^cgB51px~$_Td#8*) 9%Ny'ظwˋ5=!0LGS:I=ih]JoEjs^npyՃOoIfarhYYzqZ(^S(EP`Tm^:rŒF]kqbN+tG>]\v|E +;ϣݩo9|Oǔ1&^.VcJ6.qdC;q7ɇ6*8/]CBi<&!ߝ|3Yݼ>ЮFNMn:S8 nmv7ל]ˊR4^O0 esNY4!(Ks'Y8{5J4Q1܇F7u>|f-$XK$HD3}4.g˦擛/E6Ց@w["S@P\G%No0ZwZniݫrpYdL@%Cp)w 5ͨ0amq{U]7>,sL$-sB]{q2]P0N*Dsid$O^z\QҺCh 1x_פ+%Y{] Mt;0lρfh`=B~Y2wGcrᰨMs\4v5K;Fc;#xiG*x fvn00/t!9 KտAO;8\VnFLiN;?Ɣy:Z bFQP8ۓ*y똵iW4683`|Ѩՠ8 @llzg+ؖFҪMd _uSzj(=wfȩ$ɕ'qoo2#Icۭ?J+: po1Tc0@_pq .̵ Ӯ欲,7ZbN ;W|vHpQ=/\D[Pv,7gW^{ihD&Me;SM߉o:+u{u3 `P?aHSY̟t+ọ@ʸP~SBaj _TxtQj:e5u l1SI珊%a5q#aZW@7NS` n~wh~b8腤j8*M+O;?WA9Ksw%w 0^P!g8OfMŻHjqjuS>{%)à z#IM蝻F,z}tx ͶvC㙹|GT{A#vRlJ h):7_{_S;aMv%Y/̷Oa%^3L+pE\ۃ\?xw5,I?^[{l@{oDaU:8Qy3;~olo6t@|iipKK7ڍc<.A]}kzRp jD$Ey!.' 2fw*Ρ[Ph |QV|uKZf+0}d;YM$w^?靎FL)S9G4q$II\r)d5Yi40Eڗv O:ܕ%4ZWۛTcd^z/O>ܬ i'@KKVҪ<]!)`JE.h.W 5]A5hC=fa5z Pc, v-7[^ C;uH Evƌr"MɦB,CjCY NCz+Kn*y+#-jS&mzzD"3\ ЩE<^L Vo)~ۻ j[oYAHzZV`xbOpxG ĘxiBɳ j$knU~I,՘WB FԽ@_3 b먣+-Uk_tȷI sg_1d)$Lq f$s2 >mxW2 ^a045!|7WDf KXHs@+!ycٳ٦[63'? ܭ tN~p/"Z'@sbkm42 X!G1^ydr G-?U!>ݾ=mC30d 5PUmև6ZYox ФAϭfO<3IT AYGb3!XG:fmfm隡33 ^XA^<"E>$+hFˆiF[hϭM͠V#-sl8#-AIˬ,8P)@ǖ.mjmo(;dArϐ` #A1Gi1̎cl-eYN#g8_YhP;?eO5<|fu~-9XILAo,j-@hoEqؾK"c#b%Y0<#8PȄ.]/ +pxK=gӸ]ϔa8{IN֠az#M]9BO 2%sUхNLehֱvbF:/MѪzn+g?WAӷC =)]Mew1VLwn9{JM]jqRmU5 V"neeӇkq5Eo.Op(UW>hbi ػN6WW$RV+up2mN w7ܧzI \9Ņ+mC_-ՄrjE)elTk5磥9e\{_ VQ~w;ȧ7[jQkq۲ *0ot͚d sEel\9ښ]ھ^ ծ[,o:Up7N1쇓 A?옫 QjBNv*J.BSzC]0QO_=gג_s4bO7 *\H >zMԵ /Z&4 ~iDPF$ٔ&ZRH / {+ E)7K 'L12#-BQ8kL-:fv|[ OӤR,X(%a˾S={0E=OSq8 Z\L`ƢQE˱BEJ BAHD'wɌ69EmjRkP[i&BU$<4Con V) RjLGhx_qsrIbbJPJ)K˜CTEbH/DI--3'(9=u@00l@/MNkx\b=B[^³BAX1V;͒­tȋ@:T,6'*x" 5|hKؗ$Y lz0wiLlaq\^R})X50fd:^I]Ed!WrӮ`mnlzV=Aeڄ6zf7 .5{E Z+qagCM\dTH4ʜLu<ϩ%M"-MW".uLZfr!KӬPPRYQ0"s ߕr"QtKQ:RP\,jsx@ZiFf"[S)JMd-0\H ȕt.]$49Is64pF4u[*ZB 6*-BSP)<c$u odts9mi%[ 7 R:&8 0h9SU3ۦ9\!ʾPy]RrZ<̕J EQdNsx$S<eV4p%*)I$k)vlbZۖ8;uF8Jb`BBZ^ԃD[CP,ll`o/o.?9C[w;plAh&d= d"*SCa;M ,p/dGYFZ8cnp'BJ96|hۢ20BYHmFJEH֟c5-hA,[< ^A>.1m8p:Z`j,Ȟ n@`4 m 58ΝDH3x&2Fw'qpy#HXInB'qo4OgUanjkp2{'^LE:xVnjnXn.;2G( XԈl4 LPlHb84֘,"ZxGT"$l:Z;AvjBs\dOO#Q}J5?zPq|38%β~H;tpw!{k@17'Sg7Jj쐗F2+!F $nV?_EU_g="ٟw9x} A"~`l7)+>2yd?UM78[?/n:3n`lmA%?OfG |h{{ɃB`Qc{yj.r<̩V t4ga!0oϧG ײꮰ(T+TV_{T9=lSdU<츧bz=6ll;Vp/As%+aPs%$HEMOX*k!u`:YAXd0z70^{Gxssy:߯k2HŠ 㢖&>u&Lʭ`Y:.f,NvBGbBF/F֑Kp-V%ɯJuS1_ju۬%ԁL\d):#ր*LU|rY`-"6"F&ŠVb욖 ;↰Ap}H"אqfК*y lΚ$-CR ;JnrɰWrR&T*FfrVVe"nk4oះWeP2N\"ysI!s^Nc~‹unNS=k0(κ3,Ը-M酅.xCc̠%110~/>P Y \7ЁއCmwy>LT4}pwGapއ 4U.`ϏZ:Y{+kbCo1$l~09!6' 0Sqpudl ZxCNƕwlj[36Dnt}E'#*[C ҧXP0ÑM`\H v2).i8SPN(aDGyA2frio$V''{ |:Gle:~ =8ygue)+y5C'NGw:%56nԙ>f8F&97 0#L S^J9<6nxAL<42ݡiMSې2Ն,)^#\j%W8t*6J[ybhdECsE@BkRݝ@bk7$#u#N];Ʈ>1 >QTy];ƚL)(L+۵clG$jkT|~#z $XPyX}ˢ>$Jˮ]iCsy׮Z ݬDUp5{QǮ2E8&G=i SQ g9NrH`{.EzJL'ƹă,Dd7$\V  r}Ӥ瞻7׺ݷ]LݺYG`0M>V"\1-բR6},C(bL&*8c];8䬒)>O*T1Z}wS Bј7DیHT2CbG8\EfT$n\ Z n⤝e%h}|#y|c9Ec˞h"FuU7G\CF8 |cE9.ԊUvt?Uc?dIҕ] ֏Ǖ"!yqU}|ޫSKx\` (wEU e>|R[ k -nKoH<҉=\i"n !E&)1ZLZ&5fjY24'\)M$F} &Fp<,ڸݹe`Uw=nQIWFXg;ӇfZDD\?B*QLڂB+}N͘-Rdб 3hCѭs$Y馉, Sύ "͸-ϦVáA F|Jx, 'M֪55dBՕ\+g ȕψ &l,.( 2w¡UzwI2ѠbUNMU FOv}dI,N%su>$\p=t8^OdLV5ÕţQfG @ ۪#KϣPSzBma`rt/ BL[@+7m؅Vr֖kf `b㦫$:8:>5[dmv]d"HW^8F>oԯhT.4,Ej(|}ˎCLI޹ X.Ϸ/dEa; =_`%dѷ \X1:: &G~8=Ok?o5jt5EmDSQij>-fDQAmюyv;B tcLҮd =kdQB5kXr1ؔ!k%]Sv5\YQXy"g)߂8yU(HX'h,f]<⻪H| úč yE"ROFIUW$I{%:!eY\֢%}@}¨^< V?d1jͽzp^r G~r,8b> dq} ZY˫z jG]^U3PS%!3܆[dQPf-HZG?5"iC.Ůx3"rU1k] B!cn-`/t$;[7 dOK'+Z)Iu:ޓj=ݧ^O{Vh'5]$n@}T>R0}(>tྥ} \OgvOkʍ?~_K#8Mtm ;YQzLa;\Z>mC74h@])?swM>u5"Vv 4Sv%>b:1CuÆ,|ÆoIneiEY4 a</s-k9w {X#MkiSS x;p)bî6Yt@<ҹWėW?utݻNy6;C3>CX>yў/O1w'Ȭ.=o,Q5 8 |!YȬGjG^S;{: |Tv&92E~ V1ApWCaxX"d.B*]DW{QeZI_>i`mwMh/nn 5RpQYxb*"+ha5" Pf}%)&"n )kHT%giϴ' D +K5p!jNT#ͩ]t@XyBDv'N D:/(ZH0Tzڸ%mFYc/PKΪ9PI*M [zI̼HɼD~Vļ^^,JJ5=9 JHU_^ozOM[1ӑ3*Q\z{z^/"%T\ozO .׫ѧN_v?V8T^K iVRJ#AM# 'Ը51#U}_>}*wUU+UPwVl(U9yu s}{B/;RAvLltޚ=xi$"PHTZSJ`Fb:֩@Hs. Z*ғi KgEԃp]Yzm14H5պdh" $I 9!5#\|j%/lΗuדo-UOoi#Z0āw}gVbХN[cz"l>L'HܳhEW,X:pEؚf l1~֎ﮂI93R G5fٌ;՘x12-o_2ia?Ţr@"S]i> uIelޱYLU!OKFʑsmvy.npB65I2U9B^6Rד9F2!mFU),B,)j:`>[&5pt7lD^{Ӿ,WbH1EVY?\`$i{d$ҊFHĺȑiHT#ݻ6ńxw}ں1:7j؝-wjU#-)t=]xкH|C3^c d ؁A(bFX.'Ȳķֵ$L=&Nq!F#}pXLȏ'd.~.3R++4zк6R!HnQJy.)v\SϦ*cPSOU,r|@z;<PmdYB,tQ\Ӟ ,1a;Vfɲ,<  8 #*KXRKp{ynE)v5{8>1{\cq7 M CEb%KƦI-q!P;F~ >qӼgnd%a~{M5*$F%sӀB#E6W" Fʁ֢}pu] M +mЩ4_:nk KF1:mF MYpַcW/ C~'\+q&͚By&US()Le ;tS s>MAye#-x x\S_'T}U ^$ؑD/hwN Q)t%U04Ɓmb  Д0ȎEJƖ)d왒'vZd,ӹy1FdL9B$H1X53W3 ؄9nbкwFsS Dc: -gi9H9@ =ЈMPXq\h:4t2N0e#@#탮 ,U/Ƒ6'~ـhhCr5Um8 12V֮"lI?>F%'Vru!6bնbH9T@<D vmxV콫7@*hQ5G>x"^FJ8>tlФ^5]NWӞIaX.2RD80Ug!7,* ȞD",)!!᱕ꇬZ~(CZCԢ&.?eiC"Ad S3(!Y6RB)T:4Mv…͵=X ISP PhTf=G߀`㓸n>|Z q! N-G@mتsnƙ6PK ٟyuƐ0&)PcG0R c1`s4wV x7Iw0R(K[% v(NUC܋P,)!F-CP%|Χ_C̩vB`NR'U6d)Gԝ"yf+\V0ѧ ԹKsDlsRJEhD=xwwp)@ed))ϖ)} ʔE5w"p(2a! ;c!C옪g$d')g (CFyF]4ױ yT`+8ea8 (F Z|@@W%Ͽ_Cd|~%r'vbOp9b2RY#.tf ϚVQo.LHh#ip7'{|){iüƍ G"IYɫ:cSnt+{ 溣"gʨ2RWrel~ upW+BW r8YLafC2R|17!<>AgW%hrBJ&lx aA{X"O;oAQm. iGTT:c/Be*>kH9-2 HQa)|7tR<'lo)MOHyH1RF/z{9kzg_;l#B:n0 qz}?Hmuۇ? ԪKSBF v]:5!0/gc7﮻='~}/ҏ}sOo?ܽgtBL'\A~~;47P 0R.u?NHx}{nh\'m^nMo;:I]mwr>'^V~G7p-;r }]y@O{ :6faףrt%){zHh۷*eo<81BG.b[;>Sk{$ԅu 쯛w݇۰}Zbbrx7'ٝz0kxEL9˳rNMB#bam@6 fw6Hf]Ei#nײWd[~kKmwnKŪbSw}48\'s+hX͕AH8%9Jo@NؔS;|Tޤ,Mm\1co^0fv^!'j}m{y6Z4 >;0LLU;1\]!2;dtgvgS?rB=h>rI;28@Z,`@(\\<Ć\5E3@ETj)χ rq޵+⤽mcCƽ fqD6< %E3¨(O";0ʆW>yG29/MDhd2?t9>9>BOwCWa1ӣ6M6X)MbGݤnMk 8,պ<}oV^^ ~Y ju @0 RއR)RĄ&yz*: L~*}$zNEZ0FXUþ]y> 2Rc\zݰ҆6X)OՃhs=E!iP<2sF6,{c9V&RRbxF? DaCfG̥ Ȁ Upߞ-1gejE MUm6(4Ҁ)USM5OOPN0ǟ#ݚAG?NMN I%I&Fj*u4)F>ux$A*luK7.%Qp)ȕa%(z iR ,M DY E'*!PsDY{\]sS>8t?+IS&vq,AKWCc1G#*X-r>ySuI|„J0*1|0&3ʩP|(b"iH,6gH*cSWY!˞r,}>5ҟWAÀJ[?sQ$e .鋷?:J]~ !|uh ~Nj!| ^={ |nI[@: iEܬ*7ͪPD\҈. ZlP0s>bٻv [`rD틱n؏L-v5y+b߷E?4dGhҩ)ʐ3 FYq9v,}$io1aY4SyU:b A-O?QO\zUZj΃Ȋ@ƪWD7h*^#n1 ;L`e1m}vY 4UqU\`ڑMKdW;bh:Y0];A߱hBP.sG\yt,d.~PuZ SQ)LxmJ!&a/G'ʅ Dð:b ɇa xVO\O; #%LChِAYZ1Rc)^uud|cUM`'^WO\/8,V՚Ԫڬjv S$c?m"/?_ ~K -#Wk(%"ϿB=hAZ ZaI"*54D[0xiE а,@XST5KBTUG*&<ؘXŽN$VFc;5&E>Òt4EiUCL@'gjyqV-q>m Rq [U FgjS'6VKLEvX{t'0?zRN.& ,]N(MP2aC;!' ryZ'a sXk~9;DO]GLf7'T~+N]O,T>Eĩ +jغ֔6!1cy;i-/4h& )II3!sEcv1;?p^P]w^u8zÜ@eeMZ wrQP|BCsB^)Y!vn0 -ig*?cڡ=c뇰@͢yR!0)acxN ДynrWr]zG#aR|!P3.$OF1BXmC8;M-ݏGw5Y7j_3 )vo9-۽;0_ LZ~L f6C~_zˑ 4TH#+L,?y/J::*fلCCɘrh!)0٣rha 56VGc½8 b-1'?Ȣx'^(!-/Չ˧)*V3Eel|*x)<<ҫÃ'.eS/#5Ѹ!ځ"pkMSg0y2̫plʢp̽8kMS^VL.KK蓱fМK%"|n9䤣VNJUgq5ٹC9Nf/9cr{qbtF䚰=`R:CW|G\cr !#a;)p,*-M ,<.Hx,daf߻-@tA1a8epaEЊ_TtP5Ĕz?>#r=qiדvöFrgg^?|e.QH98N̈́ESkt$y.<_^r##L~φExl|,-&<.Q(__(.i#'d:u Q]%Eueq o'y)\vLGTdęI(N TkIŭZ T6_r%r.4"K_skhWkQR+]ms:cD @zASaZ0W/̧tv$k.IXch:F4̭乜ZjMvWs˱sq# wF~ſl݁Q}B3oxWg P'߮*ζ{w;:^*ga0V݅76bl~ Qq7.6m~ % 1@_x]˖ oVh kś:fͥ׭w7-xJ6Vޚ8ܛ_/cCTdGjfyF:(J/Kbji8Y |qlƟyF9tN\;hh#ccwj;0.&<^Swd?n ᢗhgS^YC);,M )IH4%'T`+`a .)%ÛsݟL \cַ"m>Lo ?L3ql؋T3|"An&Q $ c8ekNNJJݵUHM0EN$bc)EdfH$IJlJpc(V۔9P.mV~%T4eyA }N4֩`,7x``\j͆BViV1DTv.^Ij34faHM扐 [*5eu َCTl{m*c$YJ EM*.R=!5 %ާ 0:e9MABRaEc0f|ݥ[S MuBI%]lp/) G DRD I⡫ax~&A 렣85Jz4L h%ar>RkX!<@ZG#\#;3&crXL,Ro:+O%ubPI´R)- Vp*%rg%p9%zyByCJDإu}NVC1]leF9ͭQ|:~A߀ ϧ:[dIahɉ"X@5)&(Ld`!ڲV ܉Ƌ$@RRUX',0=j\ZJQ؂A ф*Tbk Ld"X޵\ٿB!YDV=ccH|YL0g$XvbsIQ%‪`yֽtש,XJ!B Q6TeE0^F Fcs`LOd#j,qkV`Qn|`ZC*eq8V䅶{ i@dC&ė9$ƐG-S%ژ[UhoR8QֽfƸ0PZgKfg19* /CXb`Vr1kP s1}!G\V \rlsEΡd0( dAX|.wv`+ y XY ] * +$()&A&< q8'X$+#$PꋐAV4~J,u΢,Xe &5Ke&fP57.WR̨LH20[vur`5B hH}%i*UCI$#sHx`1gG^' sG8_@A.-Mɷ%;cr P b, ~+XEeҾJ"L+%ho.Ab8# Fti7 `i#K"y{D)$h~E\!H<+Qn6uֈj^YHIBR՗%eNFT0^o($Ɨ"cDii"2J} Z neXy劵&AB W])4aȴ(Ʋ" U9۠ݎbD\,|cVɣ8ipRT"_Ha:I !1<dǀ2݀m/l`]ڍܲMNuMB&Ip<f.>b‘,}]ധhAR0Rn҃ÍdHgP j :".!ȃcF>@ "5-*JeVJAJ5oh{ ċ *:v`:GMۀ8ڂ UgdA X?<9UF,qIT16ɉ䳦`k6֜_Xud1q ,T#i%Rh[VA*Jn޲({CEX]gZzB۫-Yh/I?ߦJP0 q.[ih&΃ژWIE3+1J^o:U C ; =PɢRS,QMDL z ׁ,W (QL^#48\F넙3Z |"&T=3)VR ud-8"TSEcu VΏu~\U9-qfFKƨ85Ci#b  :rF9ae҂P/jC =NčrƇJFxW0j:Czj%bٖ\0Wҭ_F<p4k96{ِCL*Vn: b!] 1 4&YŔpIDa>uAb} \NPklT~@jOWZT;Xd3BG&^m!􇋺]z)fAJrH(f}UGƤ璷v~ZSRl(LZ}ˇ5X/[,St$uOx\~فcnZ]}ywis^^I'c[?i>n_mY) n,xkCuyQG Ȩs?Ftlh|F4 οQGC81:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:èj:Zdi7*ՍQȣWoJQѨCΓFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFayF&K'3|?FT\ǨVuYοQUaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFWcj; 6O|/7?]zX'\x`tz\M'z_jZ:)_|qvl i}g-^UVuyzz`Xu |5 bj㗋wb}[O/5v}V뗾e~6;] @/9z43$d;ͰOc.0e=A?/>-EXB:9Υ(#W_t7כW{1 _47kX{ڭPG``@tP ^-N~O}z=x}`\om7n dNΗ7~Ɍ߿Qao!cŹUFRKj6T͗R%]?/E Kg['k68돧WFll3hκtRǣr/+HNW'I?ѲS'\rsc?xxU偨ؾJQY'] ,rpNs^7%ۏ8]>vT&X;kHBu6GX/1Yedco܎`! t'`N-?Yr:pl[{ku\+k~G ;;Xdźړ3d+O׿lMOֻmYl޹dן՗WӋ|YxE/iRvxw 켇<$Xx+JvV Ɋ;, Xc{Ɋ;k;%uYaa320Š޸%#X ֜Y#;Jum|p{Kf;|:k$2P{Y׬ ^j#;jGꭐ\v+% *E'`4šN4zk&LZ j X' XÂ;o}'`R{ٶsJk:Kā`YtFUlYWൂU[ԳV96/B:)m姶)Ng?sZ讧 Κ`\]>˴ ߈Piq%[`iwRY|'`U;&:W]O ^ m~{ ;Z)VuZed&`#X~+4+:+!Y:Njm$2H,kk{Xq/1k۶1#]/1I[py+H%:+ЦgUpb-;KԚErW$,?}?My D׉o`ehoI1{9L}"ZgEԪͻu7ش SCkUqUp29~c]xk3x__?C[eX}J~Χb9OW8m@mk~ :}&c0ß:apvLa6=ti R:!ηU8;^-,0=|:-y b#t op̷iAbgX?o]#5Y_6n{&j')ĽX"|F(\Ms6Iߨ%kkJe7w W4FHEszǰ8H5^o̱v9kV'. ︈+9NnO _"W] Z+vHt: J}Ӛwk#j\&@5f^=҈ngOdjjMVDXl6MdkSc;6j~#M +ZdްZ,ڝsVlCYUkPjE6!}?՝[Tݽ6c6)ga"X,a K"J%Sɩ # lFICrd۲RLP/%*.DE*FJ] 5hZoYZ@NOTsT+sQic+ )*@#KPcN"4=C,=!de-rI՚5%Q) ִ^ቈf?juZOK9emnq"jk eqikk5G&M*#: JX[U+f.kBcVVgkl8 \~%\-<[oHuS3ieuyRYT5δuOH>"#?;8$ v%ȋc}1.0 KlAly=MK5M l5OgeF/#RO  {ͳ4e8oZ@TG*Y7\mŭAn$]Q@_!Q'WԚ+s.Ǟ昱Ƒ֑ ~|[+f7 %VmФ|jH)"jK+ ֈВjX']5/Vj!G]1Ψ|W!'c#ZZal1K(c)6 O7Mv4 dGٞF >B.聼aF|YS8Paη|rk]A)Ҹأt@<+x-ad\R6: "ˮFpyodT 0XRҺl 34`:jobE,%R92rYj0ª-S;a"=dH_(4:+(Ht|A*A-MϽ)bKR[>XԠYH8ܰpնYqr?>cA! db9WTU/4A}4QݭڏX0/US ,<W!ԃlF!geD;˄}L|㢾~*~uc-Kbc.Dn o3bBJǀ 3=B\88n:@6+MX}1[Mu <4<#yp^,_nU( ܭa$JE2Xe(uwLtXp֍|x0|dX֨Vx$ ;< c}bQՅU,TG7>􄹊uG;::VC]֫$Q; 3X`/d."O֮2X#W^`> _{`#A˘=R\wz&-rj"! 5y]%B*qaFX X0 E=b1=Y;9ۚBΨhv5+ޱaZ DA5L/a7XXVvve)#Xx;0iuj|#\Έߔlc+nϷM܈Sg 3ӲӪ*QZ|M#jw7djތG,*Wa1=TV( 85Rc9QSZ8׎a#VUGdP5:k4B*xA(mJ]knhGoMuĽPE;ΊLgr6ehm6ZY5S`Zoĩ )r^l̼_|n||j}q|L{sc$j6x c׈f=+[{ 6؃ a}wV,:>5u k昴58)kn̘`pjPB֫ڌ4IMG <%2`O~rXj9lz—g +7<'+_GTp=ơt,MzDG͏%T}f=.`>?nh峋DvRڡN#!z7YpY0jr'pYTQc$tGIՍ0| VuM\ոmlmT19\Ǥz5< 1+(wީHnl:53Rת'_630vT VQTXGpmKȞuAц-ި!@hŨuքJ!#|^iraن҂3 _B2~.O <%7aADgk|Tɘ8IZ{m(=mDB%cPJ+ ­? |DEyb$-jpۥ1D,0XKEv,* ZI2VIA |BeN V3V $!K#քr:WS;~b1CP޹}@D(GgqpՋ7B=p x#$~azo4{hw¢ҟ|@q;jx \ |;.a/0ox_~9R6]?j~wtWGEwjpupo{wzu8y6R"4?OVN~vvk_N^xӿ &=|Su 8y@}'zsTށ@h#:#:#:#:#:#:#:#:#:#:#:dA4}J ;QG>:1;K>'|rzЕM\^.VM~5Z5TI:oX9qO:7oǻ7_8J?EpI^;L"6w\.6@1L"6Ҧ#ߺH,)h:*i&j}Xp$b]ԁfYN |4h6ƘI|edS$c]`zX}xXq&}2 l2.F$b)dhZ,bCLf1)Kٔ4m[ݒY(\d¶wd!V'khLn\֘eg3C ԃ) yԐOg'~7} z!W[?F~c'8珔=QjQ{So;:O_~=꬞R޹Ddcl)7ENSr#FG 5,Aݸj P̧)v~ϷGk{\cp5[{{W>w'ݜ}:w__EnoWׇWW7g_c,b?)X|<ה_wO? WL:f;T*QSTC2gώo[<KW\DʥI2&|T/KvWL$b8ͺ0d7FZTF{5Xm QD1veZ㶎:CaMweB>??!|ƣGMx\kݿ8C?ReOm==m=IsZr⍴봭bMٶO8#FKzi[/mmҶ^Kzi[/mmҶ^Kzi[/mmҶ^Kzi[/mmҶ^Kzi[/mmҶi[!@~xu כi@ Oԁ;( :љPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPg+P'#@vjPi@@<w\ :^I@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@u*?`ԁ4jPz$ηPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@'|6}5x}v}s_#7 ?(k6x'yww VO"6D&`¶#FU6N"Xv6Xk $buaEm,#-9g'G;X$b6ƅb#2Xz.QxGDqJYrB2z1\2Xr+?X[g9, Q60ҦbBR9oDvZ;XI$b a8\,`Y֟wxDԜϸ'woB7݇Û/NW_]ڟ76 zb27%c;4Hd0J-˯Tt4D$b $b-Q1Σ:& hD FL"S~i(zo'7:5DmC^(V#%q].V{啝Dȶz$by4#K&edw.ITvg.YBW,k.l5r|Г5II& fJIb}VDlFSDl ޵Ƒ#_}Z >\ٝze0-% =̿,t*Y#W7aJ's2yd,1va/+\`fp}kz`KLVi!] u׫Ynu:zqms`:0 <?g/' {a4"('j!Ԙlg5gmmO̵`}''miB}ᇍ'(w XKƞ(ۛI- m=[0y-jӋV.յÛSCFztu/WζBw`@}ZZhEEWQQӋ1aq^=Ӊ['۟.W˃|_dm8Cù>">|:MmFv{645d7oS%\S#{-vC)WuL%'-;Fz~S/Oy#>?Cy1 +Ț՞9ؗC~G('`u Κ=kbtU*ĬSi_+_;XVBLl x;P%4|uӳ7@۽3O̓Ey7}xAsu3~-?Pܠs?Gν2 drMbwgA+6ם~'i>ݤ3>2]lU !oO\A'wΎ׫+)zol?}>^* X2O3G7;6Ra?wc)M>޻̈́~&jN0S`s6BZ!y7 #UvxT}sb^ ܢܯޟ1,3 $ϋo6;?ql>^xh~g\ć'z7G. $dl$"BoIGswtU'w݇}MYo%ć۹xo7=Th[虁۲192I9zɚŹNZ풧vV˭PD+3-;鉥欝mH#7m6e{iԚ*ckWr-Q[۴i]S:"貗h1b$@.TsԻsfCSZ+tY;`RiR=FI!Sm;cw̸]#19Q.vآZ䔔^^{G"Z#mޮ֧ՙS֕*F)8w)M":IFܥ-+)?}4\h*G9YVv/ ߨкG=QWdf?3y/RiOYhSL-#Ynj(Ȓ1'C6J^.S>7g!hUEj[oj*uvN#ϩRRJz6=\:'t QwF/ۂD:F7MϴNl -BGYG7( cm f> }9li E;g) e U2Eb3L$mqҳ 5jgb>iiѺX)Vr ` Q]8EC:qE+ܔ3sf(R>voJ:jv˶ݠTGhOMCw եfQ#2LڨĨp1"S.GO:SNbjb +p-An3x0`6ܪC:} ΅0tı&6VW}m8QE0. lv^:v95b#h< ]|$ ʕܿݠݎbF\1688(ƄUllP;L'!+pt S݀m/խm_Mz|vˤ ?1MY 3 xHu`Ky@u,}rrW263(]I TQ4H 0 F*]kh< {KHcIA'vKEvTn /} ,T/dAX?Q "bΩf # ld-Bb$ S`M1~v?>nW:41`Trrѡ",'XkxhD0(c cP.5<߂)M z΁6DR\@C 9LuP P9 Q cm2mTR+u @y@ 1^ABqk۝mg<Y[3proSGj:h .ܔ#ĕcQֽ!tM 35Yٗ@XZeЃRA*qvPD;SZ5`fnq %bl%VŁ<_ Z.:&&)jd4RYIBuw6HV۵GoU5",XPhQ *\#/^q4 J `J ڪ.<qA-[D ;ƾ8d ?ns+A֦Sk (\Bb~KEG3 +.6eyBaȢlnAGvnGU],W!BmT6Z\ǨZ I7&D37+; ZĦULm<3i5 4(R.#TK!OHהQ痽z&D#P@vPHy VMFZT!-\8EOzȕs!5=?H(Aj|dNԋjOQ I9 JR0W[2xB<p4T.TS.\n(1ˡ#ؤ3|$#PO]XtQՒ0[ U kB[yB;[gd}J'&]rᛥ^E_ݮͫzfGJ ȴVltMG7IΥo_qzww~48.F: Gx{'3[+toWPwoڝuW>}dڟ.~~{6}vrA.SiprHqrg r>r+Ypk˸0:ƃUSu3Y8OV6f6\O/]O/_;^|^N+?l}, 믵o.?Z\^ .Z<* }Y7~(~_JetkCvujjjjjjjjjjjjjjjjjjjjjjj~~(+?DnP? ~_ ɫ~~~~~~~~~~~~~~~~~~~~~~~uhA?3 p@PN4CP/3KA.O/NO/* oNttu/WWXOyMBcU \ㇿsʏG^NQ͎+qNxo m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆m8߆|`_rɠݵ >|`Z]o,46~zCOt/xju~pv*n im Y;5 N$6/ٻn#W~X`,rkgw ~Yc i=nunz{n4ݷu86o]۷ZNV.^< –m7ApUUkj=;Y,-9xW[A_`|;sO2@lzێ1EMt#v1_Lmyv-v?yɫ;oo)tγb; rDvx~Kۡd9J C`-`a[5&p)JZo-F@[߭=KI.bhL62a.p/mi~<{VwπW&/k?hW[R WE\=9C<B%W?Lvcx4; k>}l#O'0v#ݒ6-쯧7oW^#{*?WdƯ/{_9A=[u{d] =s6CNS^ޫ?zZ'Ge8'XS%,,DZ7x)XgHJGenߜ?_MiS(O'[m19VYK8LK ^}|~%`UL\HH 'U?CA] Eb>b;f0t?0{dv\Ul }82b; /矊mmiaO ^4YeŅHJ ˴c-Bz$?alDXecp`` 9=/ ߶_N7Ҳ~&9qj· Ku<]~xW^WG1n!&w!Gq)+,,ll=-8 0{%D1.,EGǕq%`mTf!`%s\{XR,cZJ]&,l4ꣂ= 3֣flFYXkVsƅ>P87>]X5*I]/{eT4j,S »__mO^w=R`a9R>R.9 mkc.9$?|KeV& )psod!`el煀uo^e=Y蘅UB)] @(Mn!`#99P/ 9=.,ҬڸL.KCvG댅U&/Kq 9ʞ`a^ XhlxB2{s܅spXB:Hŀ 6X㎻ђu{u/e?-,[9S°2 +ne<z䥽cR* B1dBF1Diϣȿ0Bh!`IBWx֊{E`Eܱ`{ EUQd ARF`ϣ9H_@۪հφ>5iۻUfkzzv9m{ }4E |2бo $ȋC[/f4x ?(ۅf~]Vt!r݇pE=;_fsӳNcd gZ-xz>.fMl/*]\;<;Zrg;a~ۅbnp7W7/?zIb㯶p/7wWo zma$շ߿yY ;[ڗ]O~Y.N_qu㭊1۵o~@%cosVK7m3Nm/6k/ߝs80;?]M޻M(45.YOƺF:;,fcDy~e͜wo1t βPދCkO.ӛ_'{9X6_q~r/Mg!_ >> []E\j3Ux;SFw#w&/Tջ7w:Ne7s˯FHEqDr~|\a8: q^æ)ҁr9i1+y i8=@?̑4-J]|5-|gn]!/k'[1 p{7LI3&=Q1&R5J =+p\5X/gG8+"bpkMC9,55f%R֤jPرӔsR-vt5էQ269YkJ*ںp3sXU,*wɤZFgAx06ˬ[Ԛ-wLȁbkds-H"èQ;^)Me';kLlVury~jʗzwkwȨT29\SfL1 ײpTؙ=1+>66f0:"Ŵ{)Ux h3K>洟:1dZGwƌ{ GpIXa o 0rхϐ>\&Ql%vMcL?pqTwiFyW*Lv,MА qq#œ"Y /Ȓ1'Co?50jMޜ@# iJj]#9PJJ6ͫ#H9=CVG;yHú)n,ǘ(5+)J 1( fFۄ L6.5l\8U T+ ~)X$pֺkc7\$FSe,:L$RX\Ս):*͚ i0:Ss]UW3\4ߐm  !p.)Ҽ>Q k96B`\%vTZbT6L3|envBʣ}BZcܕ$X9#Ȅc@fPd6>J8'X$+ Jr (Pn3&㷒2 VP8L%@ +hW;>SfP57JN1d a. BҬA$) "<,e]7 ]qM:B霡[sfEE=̄1!6K$ `4R(pPRp 3'Ah:)9oL6v/ e,74ho@œFt0GQF UWY71(l o!AO%C*[VW>0j{VRR1J{FBfU ($+YWK( T&)<'Z*kG@I 2ta*j/P&К3&Mܠ\kt[;+f&'f]"1!EMMD!%dWlgL𼜽*>9bw[_ ]]I|%‘Qq,}rCT2%+(]I,ZCUF2|x  #Cp;~.NfӐ 7&Ј(c èw`)G7= z1΁6tckxBҘCHBmwa BLV AnKԔhD;ےܱ-PI +P(h`S`eݬ%˽eo;gqOŨ~';7dt|5E\?IrT-2S&Of=[ZTI;,fv fMw)D^y" ֮ d P^%Xfz6\DYiFc0@!(!A/Q[úbC?wjg JTCۂJփ LT0%fTi'X0uP}V!ƾ8d VF$QvƠpO 9A|?Cޠ^^ǁ/ ARP0YT1"@ZMt#HWpUO=r`-`\3öjc2*hZӪk nl[[4S@:hYXAldj`%ISJ 4(k\b}C?!]ۜ-- *9DPU +w* Y;(m@Xx8kj<ae¥5g@ B2^Ȭ?H(&5>T2'Ҭړa5[!=m@Prm0qL #2"Bi8 V]곩\,rU~\ / c64k&RpIFi:uAbٹ WKBRCZ- ػm$Ur~ a`n/P[v,{f|+J,u"dFXU,V#?]a!LK!1RK \5jR|""osq|U6Lv:!d gJ'`REGgNo߿g| %g"ـl yΰLy\Ǹ-jN ؓov6.8L!]6YtM3Wϗ @ mofƛɻy>Vʷ>p^}Gp[ʦ@*]LeL¥/>puuh5g9g~.@VГx*@?#P yDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDNDu%:@!0o NTSu?P*u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"Pd:`~V@u\Jry@R3u`MR4u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"P'u"Pd:3x6]_& Xen{:-h_aQ5DkWݤx˗l@Rw@R*3!Ipm%LQτXB7E>394c=B?+ @۴^ZUnf[̺yi[-i)?Sr'.m*l>z"- yP~@B'UI׫>wLqgU `/4Wj9RŔZSq[^rh EMoX SfhjojWt[2& 8jAeJN}e ;C`(/q rZ.f/CꬖtV\-ؓY2Tf|׋GWX0yA6h\i\/8iPwD&?XvRL liٽzз]$MSGg8C&J!m#i+Va_V!Stܶ8LW(a;E."Gx6(8MCTVSr$S4eFk<h<wIv_erc Hdl#yͳr*g۞o( r3akgؐ"vZI˨5Z 12KPƥ˪z9Ƚrl!s_}%E5O~ cpqU]2 2]-;[ɍ!k1oTKGgIs$dj@GH;b:1#۝~g}GBa< )݆`fCx pJ+#8hrZQH܎ aRv_>!z6OOk!"X0}swqZeoȧyj^tbSy$qjf$ j"YqT]E$&h5R#y5c~@ `(:S~LunR "fm@1QB)FreoJkQ0IM*U?vLbf_q҇yT0|xN?Mɭ)J@:sW[`2]͐F'V\u`G49FT 0FPvaZ9(\79(u3#ht:YV2h]\֨An>ҲP+Җl( !tDX-7ASF"F9^\wWq>a)ץʲc;p''iIb;JI< @ޖfU7!x(yVe|<prSo hΦ_\k;~d1'Pl=k7C_Da XNuQ^LW~L߶G/=)2)\MA-g0R]Xf;Gid|=zGWTXKpoIe.LTATYײy_+~ ֡Fj\Z%Tyoͷnu|PsNWͨ=)ِ~7o|]MǶ;Ek6jԫbV~\osJ?*Ty=dQA$j*]}*;0pmr@èA#%OX)pB'β[Y` o6M+&|R2H*ICЃ^g$ %Z$ Ȣ7fT{9Ӿw#q fYa6QA}d4qED%"4~ej7kșw>jFRlW'E3!Y4/ {}7|xqw(&ɃbAMR<'%Kiɥ@4wSyJb0#{'74 nvk^IuȖĿ ~{O=\ʫrztxmk6͐]>e:N1}ɯ<щ5#eXbV;#RG\T(dV0=1g> '/wU~_9C/ԗKBڮR=ۼxΣ'\iIВ7bW1.yvLcѽwOa$WjFCg517WT2?u7oЍvook#L7zt畟mWT8d% =0t!+ V7ԬxÛLp<o֥nJpN2lZgbKk5J(6,-SC*uV|(S{ w /!TaZMʀ IJ<P:)T t E$#gLg 4Q(OT+3owݐ~Fa.:q=(FQK@l1bC owݐ]:u6]|6M S6vvxs] erQ: \=}+c@ݕ1@KDj\IzbRSfهqӀ_]w¥b lذ\_<夹ŷf5"-ҽ~r\ixq/nF^! *W^ޏ/0yr\+g+6 lP}} [c;W{+1VA FYIvbvn_+ŠrG?<)M5YR)X,A,,lǡv*~w*0ϑiFN)' bR`"#30&q.D!U; +>rxŗ < #,Hbui0(wĂ ƅ{MayX,L ڼ`V経zd'|)=] Raȸ%3`1i {,|HwÊ:%߂_ U`F291"f%M-QS޲9vAŇYo!Mr|yAIBr5 nWs0v zAŇ\uv#YHe `2Kj(G`F7$@3<^btV*m1sNv^uJQ!%F:hŀbP̦'{O/Z݋63zrm\.i.[.r׵ *>#nVOkhYFAŇ\TNwg[O-FRnKFhdH(F0)8d ?^K>mܗYNFZC-J ha(Wi>.2̍F$gI ZuJay4ܦ_!O(;ޙ}. ʜQA~ρO0OP<3y46d#dM0OPƵke y)]oȑW>d{]\bw>\6wmiI#_$˴%E7X1Ub]+9}(Xyp#.C:CQO?3fX123.;ֺ )93P5|PG*ff/XWTB` ؞e2W^:%>xljX ⬱9d/pji@, 7 J%HIiފ2㚀=RC*5"r$D3 A8]mMA,F))MqEω$Z*8R,dqJV=#cT! >eFy$(RfA[ .'Fxqzƣ'"g<kZ9@1)(Skնw]> 2 e+ᨌKW n\tQt5|g>BCpp k C'ɉ "U qqP]>Rdsë˴,BB+ UDZde?A>Vr>XR .bF=%6m{68EHD<,? P( J3jeJqFҨcE\ fX2[S6*R[eY XGc:aiD&'~&3cS8ʱ>}I?Jd6b^8W)_=<2*.ՐR@“(x)y_LU <!%nH 5|ń,ϛ7'q<݀"4xwi YʂE˥+LUIN<FC +r9Zէv@;ᴮ)M.[$NhFQv <C":op^{OFtB>zx*X)(VHFdur O]9 .ķ|'JhN8(dss]>V=QO ҳ’lV ImZXǬ2%I׹aH*2|Qv`Q k p +CpWA( SF=$ԡ<="qk'|";of m{*K=RIr,7V#H3a] HJR9oy!1-QLˎ+#0=$0 5FcN vm2$CNl4z63&Se >QMzxd N5Irg6a@9-\R (rZJ` htM"72u*z.3QEsLy'fX [(M̂~BLoEh/ptjd)nETe oF[s.=DuѨ#E˄I'DaS̋B1_ %rPڰGQ-=AT3RbBӣ%AlVyYe+1XCl^czEqH%B,W!>JZƴdܟVI1^[e/vi~ۼYWR)BC,RԤ|qBr cbB\ `4RS +_+di' mrgZ7_O ##cϥ$t8Z9Ul:G'.o`xa '5@@&, ϿA6̿/Ie9v=57[3U*S<_ +ve/&\g6 [fVtUɻ uJ傸nR%òQ c9+9BdeWBEF 41kF.=s"!O4|GWc:l@u j氂}\h5uHm!xb+S9a#X^G<&{"}!ڏygcƄ xMul:s]5|h|˶& bB= cZ|`jRtbU5eVۭ_T GnC嚟$sU%|\N~T3L!3Y7 ,$LsMD8!N70>j>VF}ӃɩG@.mMk5p:g[ 1__615=VX )L-h~T&쀶BӡAd3S9!dx_ש"Yu{@%0dz54cezxd\,nm*sv)P>"DB#` E撫]JUq!y>P/z>V6&KOq3A/dtOurqxB\VJQJBU-n(Wց<.&tawA0Ov#Qce;a]zq^m5F9Gsv y#j{xd\^ =nC;ّeS)vDiU3gv\/70^stUz`b> 3Y/eGtPPA e* αM=V9#GA%3SS3jXyr]e:?xVu %r0ZVXI݂ &LNVxŽpbF )f]fP~i};02,7v4s>τs2 $sL 5@8}Fd4TOjaT0@d)E¿] dFE&zx0j:%^VhhRҢ+.TitPNJԨ) UUƃ Ti/ӂNuPFo}rD=cˍ͙5"_Ϥ7fF  =8j6-| \7!3BlQ^'3,e9_aCLoT,ɇBɗ!&-6Ey"D3W*{K`L cE bT֢[q LڇLĀC#*NJ|/Wz#IW,am#h[1./yXbZ ]BHa'OKy\q)|toܶrO KY6݂?9$ ۛM:Œvuvpng^-߭6W$,rմ ߜHHppu3]u෶g_Mf QzHE#%_b Vq}@_{&MZAaVn߶گx`*m꧴`r6͎/b%SM7߷o|j۴?oZPIt_6>*϶nᏃt ,30n^,7~9|R!?Y˴.K_p)L,S+d J,\BiwxOsCÖnzg`؍*>[-6˛>9?_Ƴ_7u8[3U2e\0.;ۜAsf _j[v?)w[Ϊ}κv cfS*u=<{}/Gީ$~ݜon{'9Zlcy5Y~$.`ͽ\ͦYN?hX)XxAm?j:^&c95 o Q@%,V+|dzAgw]VWgۯ6 MttrOEeۮ%U hg@_[8GW7we$BvYg~Ld p}Y-3yO-I{//FnmZ @DOV*.{# r[8⬣৓pܐ^.^/e<Ű9i/ӬM ψ45^L؎LGIm /ftV,t?,\Xh',)qJ9ƒ"x-9xmte0 8xLUmJpfEBǨ%,E#B_"o#<9L4IӄJ P;{Rch.W%'\ʔ'**_ʅ-_z~z?^ϰ7GZd9^Kc|:%46rAc]RG4>tƮSlݤSdk^.f6ߜnY.M'߱`K)_}xæo촟[<؋o~%/f9!!N.V29sQ&=< MX)̻)X/~31eE"j8Ҍ4W)l-"%Oh?=B;n%E`>x\͙!x`*RJgȴL( [o뛋Ų!⮛86j: DR`|U<NCfhHTB֋:9 ??N 9(Dӗ l*ELxrjTjh v-㷉"(a[X0 gHEuS&]OLmz'# 0"\ɋE/r{س:s5Wl MPo }/^Ћ= X<ᓇE!r}b֡ 7>H_n9Nx/G,0zUAMڟ":W6w_M.)m-~y/KivYWM%2hqS=GeM~$es~/{̀T'22`[8r 1 r]Io+r+<x Av$}|,ϒsϯ٭V˒Z*y(܍۟XfX!34vC1Ba:ȏvmpǤdY t{o(% fh5C"_+<=Зퟯ&_<؇%:/@Zۭg.7V`K <;˄|\uc=|ۺ"M%Q"HHTIz>0Z ;^KBo#ExN8&S\I)C(d"z,=J@o( FjMē(yoXz-QDA`D6H/kK%,~ y|Xt7i9+Z s2ܡLx(M&Wx-aTTgd87R/P$ݾyYDž_Zrkۗ Ep8<'! xW'D2HIOpUU?#8{&X[D7Ord/T>'|]e4ۈZFM,#<5rw!`iBsRHNJHb1)gyN|yۋMJӜՁX~XkBmDÏFE00!35EԂkVW̝'A:%ˢ./Դ# T(yЙNp^ _ 4fb#( Hn| =ƹCEV*u;쬔;m,բ3:C.w\~E&Iaigα,5>R-]qx"_uxz") |g/}ٽSp  їЛ|Izl8e5_Kw?IyEXլ 5fiC*fucB%`ɹ$7>&7diqVlrT:KX暫D'mǹ{3Lf^\oQce`[:}}&*$I=p9,z|; GJ2[S S X.BOk~!J=un@F(dBn` Ǚ}8^1#th#Qxw?+wq0 ', X-wZ|E!_GT!Y[۟ f 4)l_qV֡}>HK7k?^8|{g1+^UH:LkljcMcwM9E 2$:[kK1Nl )>4KJeWY>(k[ƖgG4j@g B"veZu?֣ѓ2=,#A|OzOѶ2u΄),v~; Oh=S 0tO l4 [hiX{j / HOg]2:;ȍ(F 08EgD ZJЂ"Q7j#N3,*L̲6hcKXg)W{][ 9?2N=*7 t7ԛ(IMmZLdή(Xm8R*xF[EvjCS@2U'JFL+&vTS)K;M/qfe?os/JRqh"Q)f@ro*tO`ha6HT!mP9^ (0a%@ˇ 1qԴ9nQk~ dF!@AuSZ*d#4z=a6/gL%Ápȴ/l(v P/iMӿXWm|nw¿Q"J>V*%|DS[(iJ6dfE:m'{KoW~5TZc>&1A/)%h޶ $ ƳT~(N TWqG^tNYhb9ZN'ͫ cU8%B3^Eh~L%4|ڪԌ,[zz*b)W&_~fy}Xտ.w"LJzvտWo]r$xO,nEQᏙqR/Ϸ\dcy@O~q0j]4j&N0cFX&aRMl|dg9S,iw ӥ<|Z3u~ n1F~]&ō8bol [TU#55yS Z e~?̨z,Л $ `h|`afDD8/o:^k2_fcyƊ6ȊS6#<U'EyrHّ:ZLBt>zBt*m%EOrI5xA颎LU %*ͰBy* ֑+Dg)LZ M:N㨹ƊBEttqD˘I<zGϭ1t@72ڰ=VMU(06ghC풷o;F#"bᬜ?ҙm^df{_b ߣzt;&z6ZHinK:6t3Zb%$Xd)ӈ!1'$Q8Βb9+yȵ܄-M  `[H)ԑ!ZSCј%VX%m**?ʎp}@7zS  ѳ$LSDsSY!l(ڣ56,Cߗ бsqal]@ӈ"?xFwKk~~sQ,RgŭyR4%a>{Gsmڏ 1r֝j9!Y-!diUx XWx@w~peF =ȷ# Ld$F`Z9k;dݭ$ 2T{qVxҢaxEXk0*#M=5(mۨ'64$S?`cd@o鞦L#-rReKs$eEL8a\Ot<>y%זOV>sE:Q)5# Cn6 g[V+2DZ&Rqx)tzAl9@ȊZ>,ykOی`d@ay֐XC ^ ?Ɨs=@w{#?nFFk=F _)!v͔*Z6S`:ȥ𒞞n⑑loQޒo.Q^x7_%7vްD3cT7/^5tgZm)_b?UaӱwaΚTBTZ$&dndZ7 Ǵ{ܦm[m0vT5'9ܹeyϙ\ɭ&?z#ƒ>ݶG53/(.  r:ieqrV?ݦ—h6LCL|/dAZJD%]'$(j:\%U/uH,,,fssdhŢ<>/d'ЯYAߩiě xK֏E69*)Xq ]jߎHt!Hwt{CF }ҷr;}|t<  VM:;ھsN=>4 | ]Fsk=~[;hl49A f]Ȍ@ #U]U5GU>ݗ[Dm&m C%:D:ξ7:WOh =|;GfRU&׎jJ7N+hR-;DS{Od)"'UVOSX5 SyBFԹ7#rB(H7ъ!HFqZ|.gBmCd~&zЛs0h`81礂.10͆gtj:ò}r '{u]Q#Nimeڔ=ozOZF)JoHkr!M5\D>0A@X/-o G#@>k/V57Nyg0 y iWm#D 2#Q#ޑXyՋoeWC%C9 grW E҂j JʔXv&Y4nmRs{OKA5*x7\4hIkF#c.8vryqݾL9n٪djY;CR*mq3e -tLiޖ}tcZ e>^-wXhz1QcԂIA}G$2rT/eV6$!lowe{=K$?L.y&wWX&y=Y2s'c[aC-}VYh΢a{gq%̓!02 Y$ihrZ$ "5 _~ŵ]'uWO =RLb*ބ$Qݞ* قwXmBC<8 XD_M -ygXM`:1+8\rB18EIbRbM~3&1J!'%D`|[r9_~Պvti̍Q:qZs_0l#I ;v[pr%7Wj8Ϻ3.[L(B8΄4єȄ/4y1^=铵05 aYksyOf_ >b´GĤDbL0:Y C!@o[&6Ҽ:_`gmef諚򙹣E. bZ˥z:s򳽲JN,cz CYȣΏ"J#%'W*z@~)k U/:ycdDc ]r|/N}*d>®/|uPro$Z3-n &)TX6^Av?m̌g"̍F$'a,9XԤEKnzz}rm[ùTV˸Yn\%Јv̤,IwyYɟ˞%@ Jww/>4tA4~ayT8kHY`rŔn}B\] )42R$-e 6d>kBTɛi2}L9t&8>48K;q/5hġ|p焂0.Ӆé@"JAΉkGXڮlޢ)WI{VqN Cp1nf.\@葔Ҋ$+T(8?AkyA|DBg+}o:ukrM\#pvL~>ٴG$]B~V|n.Ї% Ҕ]v!HF糌SMBPpC8 i1>0<(/uCF/A-b()|AvǔKf8jl]`bGa5aA˞_ E`)3eTtġft(Kg>gzE砰 z'ArPFeElq L?W"8)(#@ܢHJƪܭ5#׽sQuif9cXVW]vfJP)s"z>[ @7- 㧟I'w#ռGHV2ނ}eWBZ*HLkB<.hME);c:/ F&"N4~,->>^:~Y(1LT*0AT $CmaVnh!CEzp]$YQ KO)j1SO"g蜣'h^4I9ρj+@L_2NwH.g2kGQ0P/SΞ@Q0셍jkKެzN`Ig㳿~PRVXП4. t}zPjKD&6;|Vk^T&C~LSJjO+H  t c)G1xNYqg&ÏLXJ ͚q~dF.-'ٷ28!4:`[jR[*i$ٷ?qR/?Lrz3ytVNz2/Z7+Y/雓S0]S}=wfncL鿚*S|F9F&K8RD a{%YP;.(\8F(%Er#K=AZyb|# B|}=#1mq2;c`;0c`â%u@*8ch0Zp&BV q3&ZCo=Hwpre8&N<%bBu@`A5 Sr'ڂq:`^/g JÂ+bb7Fg0(ez<^%Z '{ޤtKn4,tSMo䙁ۀM̑յmbȒT#csN8 UP]VrP#}0gppDU &W)AqڪBcrd a:_[ ~>Q] ŗLtrNǘ!Ҍ\.e$h,Ab0)(V8bL=X!'&+ t[0? 2y0}RZb23t@hrkϜH*㠏Q V\pޫOa69FVSǑԠ<)qNBŋCwraQ1)8.kT\^&>҂;0(Ltʽ=c`>}=mgC {h@JMQ h` |~c`u [%KVQ|:c 5:ޙc& 響6c;49zӇk}UP?U-&'-;?d` )t FN8&N/>*s10qzY5F( Ǜ3[r72=yT˸< ,|w𽝖'*]ҩQF%xqc%H=aqFl)~ؒK&fGǘXM5q/0ҲǴF ߒ%[zǠ1e㚶Tm$'cj-g28z1kwO~=zq{z|h8ag'j.{,YzQZ3 a}`8DDCeRv&Bi<.XjgmH6nnY`?&~MVLSgf~դHQ)5n08U]]]U]9IFn"u缪 m3YYr '[Ħ"2„00Vӣ)h/W|\)%qahp1"I+@F` pkB]XxI<,I˃Sy4"@20q4$c*6S0)hܹ'[?hNa8TC%15?q/XcfGBE$>=1 erE0R?"JmKEYX4x@O#"Uf"tUDwCqJo&Y78>/w#4yq0K;a^=܍hFe70qX]&ɇAR>0t#CպLb =O`a؋hTiƑ_H2L/*Sz3&4{E_%Gn_tH^-Es*]ag~|Y9b0KA笢N|o=~ 䚀Ac_aoˢў杛J?U_i׷׿ >^NɮH!=qdahqUqkhٯ4xH۝!E^ENaJzR͑m%T R<ȭsc& 2;^l:(h{eyU"Yex߿R |Qm"G pp/qlBH˓ƎIbJg y?!1\jzѢVY*?O؏>XR~zŗ8Ra0˷tjŽ JYQqfEOqTzC N^z;9H~mg_+Y,[~{ X:vWsw7PUU,ՖQ| W3vl.}O@(th 9(!A)E#"4Ly,@<]O5N @9xsKhG6hTBWM٢?eqp[-opKW2:t)+gD}Y%Ћ_XPG^qm r%??eQ׼G=e?h_ݹ^˿lۮFϊX\t .YyV0뛾R۱2ea~._RӰY7 62Wem7s #GU{#!~{UF{Uv[8 z?+UpJ *?)q#eUr` Xl.Va!kΩ@n94[>c3U*5*@bP0-Dq?e N7e##lBR]+db=zL|-34a,2j=Rx˝7I!e{uog&ʍx>=S)0}|r9u8GI d"P&4q)e!vB O&`qs;$sC6spڥ]ѳ~Y Dyl{=NZy H=i\UPNbګywQimǟܯ̇pM L"*#~p0]D+Ǔ\5$NκJjJ99j֭ZgnޚnaG< o:6UzwvM YA|jE*|qE[ƬsjA1fS[6ҲpPy %$@z5o8='*\`O0'YNL(BFft[u&rEuei͗ `}ACױ]&F 0O6C=O2K=oCU2? /37յ`_qKV{RY|6ׇجէ5΄9f)Hrb*H D{b7AsOq #9nq7Z}@{ذ`6EaqSMot>A.[mq7"ĉP$]T`RJ O| {̦1G!Ob|@\ngF6-=/V۵tkF@FFؽFEUڒ3Pꐫt%d*eeo q1zӭ2ZWU. K~~i+{arMjDok%7oJҨmua=vbE.9I';_cʇ1`vLg ~բxזFF[v6L1s8C`ݓM2I>Lz/YZ.@_Ea!shЬv bYx\|SBXIp̪ƁHZC^ lhEȎ5eE@I|aaߵ)P_ac{ MǧοmB $NV=|Mc@*@-¸Ȉ!>%sc٦Cu8hT4 @'vn۲DʰзiE^/ l?"7sf{13%S 5~ͭD~Bz~H]+JL22U5Wp90m`Q( isynaJ/^zϯ0tUĦ!ķ݈~5 Lesi#*z9eMag*>AU@` _IȷpH"?w~Z_U٥>ܽ஀R a;Ut]2z3Vg{ߺ~joo+@{7)Q(GCl~ck_gS=kҺgyܗ亚 ?;TO#?tsd^2SݻyVLr>: afv-&;sw|$90{S/w{dǦPdmb>h]MzO"}T;RW~؃.@#_16{]ݾY߮nsGHn.ii2۠$Qv_@gnI̢QXcSyߴi*9xpOu01R%V`Df%~?&D}qE3@a'nZ%7 c,wNyf̮4.QGSҎv+죸#-y}DDk0(O_ԓaBQp\\X EԨZ~qoB?]dqS'z1\23+"|24]pG<#x jڨ~u[-ׯz\qF|-q\ c\٫~K *q'fu_wrANd؅8 ~QDb;} c0n 교/Zj}&]Mx0Aͱcnq1q3l*cm&4C_ vBgNR d̳?]\gOd;vk-_Z{UU6vyNa&nBPXGVwY)J2UKJjWߖ@,Y EjmZ4(EƁ|z5#=S -8^[?ZP#+Pm}5GK)H NE#Xw dD{nK<Č5%GM|JwBa5 ]cP:,Yt 9f<,k,_8"lu_b* i#r) `YQd?eRw,K47Vr\N՛ Mpi%oi\e<+cBbJ󅰩dO P2]ނŜف[(m& j*]=:*xԼG d]Wz\Uy\[KMjZFIaqဴ~wX,/f &09F#9"egjz2e!dꪪˋK'7 'I[)[2p#;a>=GNSؔ>3-=.<XgeMO?kOp ʰLjgJ#fAJXp=9fNmNs@w!UdӠU}Sa'CP=CKȳu"WNQb@`t`.ZL7u |`wE^^mhB-`78 qȔtX{gaLaRd´ bW4B1iJm ضh~lEq#oCL(,12`*@=Νs>\2%2-g  L:YD c '˧0F!Kb<ωF=Nh#v#6,D ѕG- +ցtWqlt2a]"j#݋t:.O)Āfɂ1נDI,΢Өa_ c1愅hL9[HxC:_ +zGčMb:YДH*9)E@9!e AI'e5AhA. )OdLb6?ZA1`pcGp#4BQ)Kثw눰FeBX:g3 ʫ p#($ 8#1@!+ > ,YJ`d΀btdZ2œFd(h'Jp  `\`0x^S;ێ YJ`#&fyR0ːV|E0gazgCɇh8t]:Vv3TDO<Ϧף # d|~.6T=8u'i{SP,i$bT_E G ΀0F~(' ֜jB1d(pGJ @ɰ˵_{'*|q!!:0$aY5VIc(0Fj-/Rd}p3b>ˑ_: Daڅ[7 shʚ( iJ' =EA k|`^Ue{U=9äowjV0N6eee{-T<}>Z1å\;4d69}FX E 7P̦ek.f p1-)Bݰ9Laݲ. Ӥo'/H@ &}V={<O0-6aaPF,[_jg/f?L߂d0c@Լe ?18675);lzh3O۾R}s37ƟݬWX"K1<^ !VqFqq7ZL穋nFE_VcC|6=ΊԾyυ˫f vTo~*; HR*>~]/]i"g=+"=.n(PʍEm2":w$sbrX+H!3j=d~n5ՋeiaME>5EEճW0wx:΁d(OpF+Ho/ 0R=ޅ騀vI"<|S1T*~mr_LS[_M?MgOyV6Gfmwms nbBY^+iq G4S-zдoŲe=Jj}?nur.>Xd=]&6D&6T,%b7_y?>3}>9ϯwX@rX0wGͬNSi6{XKV:}TL`ͨVLCFfR_-)uα=.\&q K,¢oAђk `Eg#2{TvӛD/ )Fa+߈%’pC%rrϸ] ] ԟB஽{>M:@T=Ehۣi*799)%1`Hj9j#4+/ ~G]3V6|`=L)gxfwb 祹X8m7p{T6_:J/L AQv|:+Qѯ;kHe&#Q kaj3jX$fk1hFs+%d n?ou7&z UEW~o^~\\a0WZm}۞8)b <* s~!UuY@iI(JcRfuǍbtٸ 9w#2tseB7bvv;o:5d>+)&@[)Rm7'3 ;"jI]tty}xJ-Y?m%o^GFC>O2j-$* %^F3#a:0- 1s+:;ȅ4* L1U |Ȝ ('@c}>siLbhlwDc:8XG!1z ",iP΁:KFbu kpZ6P@i}RESXDzԻZ}Ҫ=lÓϲѩKZW44.o_ s,uρͅφ1.(w;2=%PU]2ގ U{Gi@BI%Rاm1\w՗5'q3P sΦlgS*,a fh62Ƹtxn?r;k!.nۙ0Z.8rb^$3\ze>eS'%qsw'ytj} E70}?IW0 q19O)к\jǻtn2ǡy0gfn)/dE-|6;msK:ZO|ؔ-;o{vꪔq\.xk p u6/鑮L ½0owmnC2O˽eMIXX>t@[p>Ocv4hR0L#˭FtG jG"Fĺ܌J.lEN(wp)ʄ1qV{tHb:ÂSqHtfjFKIQ+TšewSx0oUҲޣ,PzCeP$I5J]4ÚĊ*q| zPwpqJGsk\e;x`y=I*08SN"`@2""&ZH0|&\c K?C= Z9Ag<2Δvs{sQNcF1ZleF͝QFGR ;Ey̅#z:" ^E;}XbAͤ(w vrO. wݯΡD6\Tܖ펟Qg^[z?PC̻3"lnyt%T8 1ؓLkKI RDpCJ,s$#E@K {Pir]ktҞ}|2be/jаZ&|;tYé0|=ܱ mW&(^(֎&py[05t[ՅII$RV .OdGҶ hn3 \-q鞗5fEwHJX%Ѿc~&,y <_"v3'%&T7yJuneB"ت3hocX+4f%81_ZfZvq2t*m4.AzUCn)C㸩v򮴨|8qx;rt }Ts~ Ta>-3Az#JqTȹQjEw$sb@ÌZ ?D ؉3o gHRuh(,z:|N?@ڵn :aP!Ռx4=(7Sog4돀jPܹMh6 uRViw ,s9Z2 -h %>Eg*Y鷀?vW$9^'ٰ/R~_Fk~>K,rXZ : oQY*],"q 3UP,zk)1gDoAs_wBmYl9ʬ#9/hf2,q $<֙a􏴚f%q3dFq :uysUɈe E9khE9fYc=h ß=cK3}92ɔߝASIz^EewiT7CB m*p -nZؤQM'TKo!"Pe#{ʘ: V[ƕ83IPؐ(ddJ[E|얄3v=c|>bK9{3aw(Eo5f>pez)8(0ϘՎb̄V*-vD :eb%í܁Bs6v ~mX6MnSn۔v3&v)Qk446:&p\pXQv`܁Žs6SDiL_G׷Uٻ6,W=Ř\< b`2B]m4%% }OubjMvS@!]O37AIɹGΉ@Es dž[JFmX(m(Ȋ/;~Mx5zlup)ٸHeW^+%Єŕ]aw[EEڜZc9$mvF`"hWpL7j"pPDpl#E'M) XLBR[ S\f<pSPa@$$J`a1UR`'C$ZiGۛÍ~>a+pW8yt4W%x[BPfy`8R0' Xcsh{Vk6=߱H.Ȅ,]䂫\pUM.XXH % Q-%N4e6N<xvQ;hyђ\e ˷Oyfz,??xs(uz=/jWW}[;M[{G A]i-ZVڤjk#TSQ;w1[)SVD~X\ b qNMd Z 0 YؑCL$<>q*ceGt1U.te2@UDmz+UU.޿ o(CoMDU+JI:M;TLPdN8 àq:i!rQ`.D p/TƄ|0ylwjvV(?Uʭ) 5Y)6e2FU(д$d ܝ,7'<ˍ%"o#8v=B􋩊I&"lFiCKL@×xx๓+ sAߋQtE9ye!2-2}$Kqg5BhdJϋw_bAe2OM:䄷˜;xc˖_Sx8W󛽚g$v6BlN dȀe+o_jp|9~Z_F*y8~f ~bJ"<{4 [Oq$_g?fi8ˊ4 9 B*OD(dBl\ _JYWaĿdx3XQ_/'+jXkה:;WF*Pf-s1_[~wd r]N[T)=l%W\K] . WT.w7ur9P-P>bܩ\;Q=" JgI`])M]5_n#8=Rۂr-ﮯgK=:U~^e՝Ot4B`2qbo󔌛IKSK&ZK4s9; [lG}+[%PuWE).Hވ"J턷.DBvpEhr"zyB6ah|9ɕI1HdlH(W)xʹBQ  QC}7с;/;^Y%#RVZRXSHP2RחdS~DD]D!qR)̵sHpS~tH B]&79_RdK}_a _і'[p<~e6ohѸ 2Wb-#. XBYN%UP`l%$Ӓ`"6ƋA=BB6v!V!/`Z-XB `(i>d*J=NZҜXAy҉&k2qImQ"|9*N$ %^JRUjXDsJ'h:OzɓE B ,Z N+) kppTXAL".hj$ pB4ۜ^ox;]0KJLRD&:8f}Ra.k"1o8[ᙴp"`l'Q*q{#O!`Khp$0ؒ^A' mbN:0N$O*i-6*$x78Eb%иHSPha6(0hJܡ!WHe;RAQ\Y / FCr\kCHEaT3ul37gxOEvN)BsJp!1p  5'IEL λzܾGþOc.qgCZ>ll(+:)m-Ӛ9P MĒ`|;c:&‘|~:]Q=oDx.0g=LWו7>d,*]}C,O [;.`׹@sۿ ,.}6׷Ry?jw~m;sk}ϑ+TFP:VԄѲEH'b6VG 5}4fpMxw3̧afial"+̰)|CgMc,T}CJ?1D`Ft!&"3,dLgJI@a}X A8SY%3<5S\ZRCr)u4 Z[ޤrZ+.fwQNԾh'7NgO͙{)(ڇn#6')6ᾠ!]r 3?D 9+-kc#e=~̻p?\F]+ <[&̠:|Z\7?o֔fOp|[KhgxyZ&OGT6F԰vㄹsp M8a)5rTv|z5{K~>Xi" |n= -Mv2D׀&g۷]+CH0d9f -Gnei x xHh,O8M4%Iv Zc<0v EOV$p݆N B&l$6dzɷ]`+âm'ߠCο1RӅjVNExU !9s uH,$Bd 6*I$nE?P[WC<ٗ/GNMl4,dG\冿f)`!XRF_`Ryɝ96k(7|LWfQ,, wZ)= )D%Tj FX&.,XocMr g$ Ӕ 52JX\1d6azK;(Hz~, BpQ;YIlP/OWoOZLE_{l-ق/قrfT}ݷXVrHwZr%TԹxZ+yɤK#\Ѿu˹G7kjЬi}iVo=zC g}k)+rP(H׽eqNQV߆VJSDyBo}mݽ@P !0g "89%Iҽ+U .ktw#.dDt~>sG '2OdD s7},>ğ`I*W\֎Dx8Dnc--c,u2񓥞tScfQjH͉"9$\9&.WOE;3dfݿOLNgY'e*DX(G{M@DЊtwѦyky GOm W#Wo9d4tۥY W"rq`Ż7h0r|6vk圛|B+9|eͺxC5_\}%SA<& Nit\g="0h~8LA R#@=Q?@t;u:S 6Jp>FGτV()5έ!9x`) L`ho Բ%F;Zuqeq1'c[+ѹتmDS@=7vd9]w\|NWo>n~"G% "%G\JL c}!I12֒3g}.r彦yby:LƲ %g".D YƩLI-jXzMtg)3f2:)E24뤳_ʲMw* zCDrY#!Dc[V[\$.iOu(EY0q'ek}̉+z~ɑ6/;Y|pp {RDp-2k)+ oGVv@¤e-"BZ1$'a= ;Գ  4W$Si`q`sK`Q@"<*% PPlfFX0gWyXB=ɉ3נ㓥nSڞ<qrë޾Y& ^^ 7b>uU6+x:h4i^9ɕI1HdlH)xʹBQ_};T4Mnt`/L'm?AG~ʏ av&wcNoYr(_p`R=<§! 7R@Yf̪U1O[u=zĊUⓕec+YtL{Dz*v+ԛ\ji|ر"d!o[enh:;'+^-L[{hE+z]ٶ= )=Z~qٺ0g A<ᑇԒwL{h5ΠMm9d9DZ9ͲS*9<褢v:nv8|;қ/K9VI:MruNZ^Q$W5;MQBɾP|= }1 (!Fُ?.J2%bq)-+-0֧JԬuq&dYce g,;?tbŕQ$w8o xkNp!r 7(IM:'G;pJ <Ȝ$^8 &x q3]]+Yd6hێ@TM'ipJKqX7E5:Z~6(朴RdxM۹3ToHN;kd4Hc,qjk%DKTx,(Ub8O29zyb( o-غqba+jdyRi)JMG>z"IiF +e ,d%Wk[F[%úTVboepUwlgďGz~vhNAL1P)YHBxE(6L3#S3 !XybPBw[ 3Reœ&ac>XWi^"8,qB(rJ9ă"x-5Ga◵?c0Oh<q D6j/oh-򏭲XFV뢳:߹D+S kESb/;GYeUoޜޜCɾૂl+|fOx7lNm$mRN{Nڢ {F .`z滋;4*fUָ̓]ѫo'V/CU'^c#CW ]iߠRK;?@h ,mKظm]h1qY! H&iu8wz΀uGFԺRCɢ6Ǜi'o >_,EORMjl~\-qؼ{(`q.^j&"-saM7%RZX" q rDE_J;B6x/kzbotet <6J0>FGτV()5έ!9x) L`hQ/H&LJ~~{G+=Cc] m.;' !6ܿFEnl|v۹{gd4ٓi_KۮsdvOٷ'}w_?W? 0h)" WmyuBA=> wfK|_Z M\?c܅I1!!|5ɦ)»wxu @/g-Ӛ `1$1w5?}2 Z]zeW$x~vknyl+:Ңgg|0?7!{; `#ѬMh:]'W%AF>fF?'>hq7_-k=f AsԵ[k[ʀھ?MѹQkwG8>|O+E`NW6{Q Aۭ#_F9}?u6>G1cd߿ S%-C~d+G$s9Zgy1'{0&ZGɠѝRCl:9h3Q#5"2Շ\!<{C5?r41%IM*Z=iAa @=(_ c!Ke bc~{ŷ!=|k# 9bQQ!.($Z'؞ScB1F9k6N+PbK]zN( aL"[c5]p 䕖JɀBTZ o=ĺ#A1iK6 3aIL^)eࠤ@NI^S"Hgz(1y$$QqYc AD* =:?ל“uB"Ռxd(21i)O~H%&o8'Op4%8xld=M ܴUgz(0y 'OA1EHH*@cDB|x&2*cPbz[0=R4 GOΉy3p@@^y e`&dȻzG{ R`ٻrWT~J1wLUvT!U^ ECFmǤ2ɣӾPg" Zeޥr=}^ܔX7`?Q#;y˔NLD{w[Z2T E MPHA-ǍkȂ"V[!a\ƣʏCyu* 9Gh3 a@K^B%`k^Հ#$P>{k$@^\d;Omي,@HĄb@z(AEMZ"ߛf%tQ3\ؠ#x協>1a]=rEPNXJ`:#V8‡Hji@WUڢS ] * ^D&e Vnk\w6929PސGogSMѐz(|0tWT6K+ klUV$eZ1CK衼S0 "FRdBz%zTa J-  =wJyͲ0.Yp ȀliDARc$P/ꭼa{-{$$L(A $(kc$P>H/{+kHF0:x+o`J8(?BBC-P}B-o! Zhg$I7FBA;$ -(+9Q'#򮹩FN`]FKj͏Ey:2Ǜ dRnz ֞11UD JK *&^DMD8'?>Z$UT*<( iG2g%c^DKK)ɮ?pFrHoy\]q'>P^C+{ϿNofO~5Ϸ5~6 Q`|{XĨQ2* Az@CMcf#3U#i =#$PFs[GCg8O y5 ؒj5ULji ֓ X4)V5vj;ftOtZ*A{! ^Xd)&վ dC}fEŰ&҂!h7.cH5h)%J-#JI;6lKCpEPkĐGH|4}`BXp|bMvHz('32weT5Y$ԂBFdeR>S.{FC7Ԥ p)Kb31J;);3=ɛا`X=B7RdsM3Q7o7[uFʚ&0R^oYn@w(v}]_{[TB@;aI&+d@ЃziDGɼJXxI=1z(zV>kU"eoQm)Y{LD3 ]wJP{#i0f1*'nO0H蠼>NB 1ϋ* Rr.aR~ʃ~2`V1"U5S\c$P^Y4(,*YEl\(67FB5ʧ̮U vNJ )"8_ :tc%/dIڈA%*uq#!vxFf0mEk[fXAd #NWw%(A[l"lMML8#Q$u))^-=&Ë6jG蘕9c$P؞aDQdgeWXWːӐ͏CFSA$xVDl]B6?FB]j +91gNFфTd2 ]5obIMAhd @ 1:(odpm׌.J'EEq tI ա2c$< 3^SIs^\]]=+_ڮ#K-wfv!n)_ݯ$YqB@ A]`s cz/ӛm)gATJ*RNP8pVta)|Z@[r|#Zܯo QE`f}Q:bu:DZQb8BakX#B!aR$D%ҺHY/_F QtTfWryvϵx?=9;y~vWz7sro6쀆[lD\v:ݍ-xl1 q2/|4<<Fʷ῕e_PEW/jJmP:ACʜRt.ob[񼔹0K{@1~B*,% BRBV\d^DA5@(m-(-&dYVC)V%dBHUyOy祮jPGB1^u*-iΟjUg[B;uݾJTAmwȟ3=Fτ^@Qb1)֛Zt=ox7ٜB;oi_pΔAj"n>m-GӶ=.ApZ{yN R@b) h "S@Ust̀0j'M=8Dz>fWyo/gQ}d>wD5dp*V+Xk=MfjVYbl)ƼLv-FdtFG ς\=T xY!k8&Z>̗ĥy8o 6kv d|%!wmH bjMΙj7˃ݮd 6_ak={" Hj8W'wÉrfw.smT-rֱRh#Ydi^($^zݲV.jU-oIv7~o/Wɨ㳽$eɈ)I'Le}9[˧Kqy%WW+Ԋ96b%_Ly%S^6'WAK:ƣ,.T ɶ1 0ʕT !1 LLmA!9#Pl e!JER5+rZA4`B.V;['݆32bNǞؾ_/PZA;.,ꙍO ;ߞ1%ަ0C*1 %aȡ P4V ѕbD m€R!{)+eFUm8g󫫛w־NYu}Z+/P+WLy%HѝݻaB8H7~;tU+*|(^I,yIۥǤ6-yy4߮2'=nOʗ _AR}ډ<1 tñ1%0c: L'= So4sjͯf3vퟘfbvCs[ rif>l?}72-w|dMĬ׫&[> :\ {UBç|/lUkp4/½DeLv"d(}2msm'_Jf햶WI'{Fߊ/x]>].d;,s200d{3dOl6&xL{#P:4Gް ;E>" J^<.]?P}l$5S?t8Ϩ4)2(ȖF!aHLPZh^XZ77q! ºi묭Td40ϬA!\pCUdACYm{s}33gr|DrN͡(k2T5NBSN3HsMBU^K'Kγ(E옹#URHhhK*8*ԥ(gRA7̦=0cp[4F392YYՆcqZW,WQ{z֍j, xnz uYAn(Jܺ;]%B.o&wsH^KG~[i;&>yָFĤ繖d_voY~me= $o}{Y@ad}e)KapK|vq)^r{H0+j#\T-]L".fRB\޽*xHo=ƫ8szy1 >Km̱,|SFE !Lĝ7zu{hrIOšvͶEZ'BZB+͏Unge %F4y^*.|AQ2ģy.xNжhӺļT)dyv6ۤ2P]jzelF}·Ayq2ԳwU]bVéQV#cq:b{?M]IJRt˝p(cvkA;%tR& @fSvMSS%Muu?91{':WUt[wllWI)%'wY~@}em]XLXg  dZ ȤM}, F3;@)C ^34wh0i:R XhewF`G'' 镘gnwމm&@Mqn\l=՟ͱEGGBG/#Z8("5Anl7&x3F\L5ȷ?_sݏgG3#'o;#7Ao8↷s~3>#)l.;y025տ?\Zr >!B!A~vtr9!.L^ ` EP&ܾOkr)rIb>EcDfޛyIGl5WMwVۏP |۞n:9O<9|}=EհyRnpۭqT=<.h$VK d\JFzI5<ȫ iCPVK|V=H{fUkRRDGpU[{/ؾP@-3"39u- B-3ڙc8qqyO7 >]| Vo-;Г{'7п c"-"Aռ07b:qaka$eQRS[** I18 d~]}~D}aAY"{ ?ƃnJǾJWZKbRbMZ蠜~xv_w?Z>=Լ)"+G >0v1yP_#7$< ؅"M?.rs_|*Wiع-]aٝc Ul BtB/?׹/LKaAWq I:n[__v9m.0 99ݙgX;f؝p®5hSj7iN^zۉ'tMQ}{]`~*X$KN6nfƙYMͥܦ wfd㉚zJ)UoQV VBCl[ov\-ZP+<#)`_ov\#k~ci^No M[}@3aY3k]1u{v#ww -m(O9l}m85*$W)$)v5D3D-%JŨ"M j˩U)lq,YぅFs!1AjjBP'(;cBALk$.NIn&;E@ϼ";u.-l Nb"vI*m)$`\zrJ4P%nQX^y(_S6g=l, xnz u 7m鮌ܺ6w]z6K\0,ؚ`f5=7fk{7S[l9۽M}5wyf~(7y8X|䥵 ηt] oiŰƚOn|[IYVO>?ɭPcilsF!K=#c&NyZU_XC!8%ǚ I*,K1L$d XTIe4Zm2=WPhR & 9n,l,K4&|52¶P$P)*5 PRf21{{ G|P@@)GlVHBN1Xk^Ӏv4)OKP`Fs02QRںlR]/ FJ0hjRMmp;TjV6iޅRow7 O,Gn?\E<e11ˌ6d=gj,3kC!)&z ,-m)22ΪdXKIVg>1M@Ӛsp(T9PWY gY/I5wg+_%K6`]Z',o67%o^>t`I*'F#&FEO"3T*Z$vzM W KYĤ/.9[=N`sb7yi8fE?s з=y(4ЦJC\)3Zh^1N0J5kq+GEKv{l^7*n*n7TA҇*&h`PL 82DQ1xΒj=+d wo0G?qarr&ܺNw(`Ϥ;3~hDe˿+2 r+Bw݅UAqfP4huY'i/{ou O<.WzjqapxsT|)&]lG.YvRcQTRխ\oCpW%FuXԼ86N <}p<\һɁ [ )b) ."<*EGyL)ƥEoL\x8&|\O z1xitZ :}yځuUrjP{ԇF=ڣM'7|4h{Fo7{/(mmA9 -wf(%u<>(y5{t6=zwI᜻!5Y<P͎s'>DAE%qE˝UލQo)_;tj177娫17Տ??{x h|?\Y `,'E[?[L\w_޼{7fŻE 5QJOr~wPr~Z,N,|hi=GR5o_߾kՓ6ǀpdےh^>R,7,q+Q+QD/%QI@dƇF\s՛5 oM{IxR$LBLT:/ . "|LӢ&YfR@; Tx㩎F&--Xi#1H o174Z|w|mWrⲄ|_\ʹ&|GMzuTػ'1uB{JrfNV6챈 JFE ˕ u@D㢶 @ U=V΋wT-8|eh@*f=0>l*Xa 4h"jȅZS rCmZ cwtj|BwGU: Sb9>cJ(p:kۗh=s҆R;SK{39S\+neB?h+;;2DH}2vtftLBeFTHu]?ꫩkHG50@ BJ~]C.pJ0|lRl:*J%wI,8 Q5"7r`8-=!(0z$P lY=ؠs>^SZ)Յ(\Ӈ,lƎp]7X|(@?YF6Lkހ%qֱ$IOJ@bQܱpXTU@"+'&WrDJy F4 Ѡ   BPƹIBgPCfO9Tzpomjkމw>] O?Pv#Xυ ԧOذvffy?}fONpH# KԮR&Qͅ^`+xcߥ~|qv+B;*H~#uϚ'ӯ?!(߸Kt[Wx/&ST1ed!ijb (w[ %QMnp]Y\)sFq􏑸rW-#;}B( IA@  C ɵHPLFrD@ƈJE289,h)D<F,9TDҗ&K 0DݬuVɝejg_l[%u19yikV8KƋxb*hpL.:0] ],O&؞#^œ/Ǹ41 FKI] ZDY4D4'=s#wIō̩QjBH-,W!}Gd GX[J؄Jf,6Úq?K.,&/ՅPX O^>g\F3Uvw;:݅틿lhy4<h"!Ӝ-f60ceƥDzbDD02_wd-z8Q %!N e.Q!:z"^"K@ օ$AEem =۾!M ZG$AKH.8!(HC=q[E9q PVVvJjJ3/M*W*򨩍9"*\(Fr~Ua<_VQ@-4"5I}/ޣM hm#x}/5ǕP@50ƁzN.D=K4 3KHΓ4ൗ UAEQC7kFd2'9Ih{C։bT`|Ty!48qcJ4SaV6 Gl#J/+J3;B#y$R(N#}Io%c>'p)eӀo?gg&=ug mk-蚩Dh׹׏z}oHl`vul{\@m{+ۿmز[g=?5>y8畖!_zzOs>F揊Zg O9Nx1붒79}͜6Ǽ9j=_77%A5bW j/u y ]j)BHo"Ԁ'q4|Ϝ7U`BXD; DtJT@A!T nVք{ej<H;nt,}Ia%_l]q6XڹO6^%Thl㼄IR4Xb*yZ9^=4WzLʒ(u W<xH ֆZ )OB3T&F@U(Ά O2Z߱6G!9AmH6>g$ED o,;:Z?xwNZRJw)762Pvh(Xj#jk%QDKbLx-*yb #p#x+Ϻ&Lia{_`!vbyjۯW765X;?~ޗ{k©1s0+*qC$.}b"F 5A)B3),C8,Yֳ"[h+@6y[~sa&'wa­K9;|t Ljc>㇦O<\"s 2+0+0/ Ϻ"t]XgP3Y^7L;DXUdSzhPdRK;<>,dmK踌..Mrפ6{Y~̞e =z,ÉЙ#I[ΑufQPCxv,Q ǁ1G̗beHqdOe,UZ1y%Ie_/e_64 wU\oT7PnE[hӇO/@B h2*"(qRt$QRa\:_υs\`OZVAӗ(H^W] O!?  yPVʾs6] "z)rM"0>4zk"*5j}M8UN]i)*ň< ]JG œA$$rLӢ$$LZ9H]Ou42QǸָDphiJ#AT1H^DSOӞg}b^{Gtt}{禎 &9i]'0OEzS>pY*'HR v#v6'Xx;%.dw&C{77ȶf>!Af17gMnnluXa)֑nk\֨Gַ[iݑ,];f~ lea~6/1ٻn$W J0 N`g`۱vKn;-^]ɒl=lQ/7?t*C?9PΊ޻E2# ~Qvvxs2[(SēI{m+!Zqk'Ƕ8Ł'j %yXl[ʶ[ިunRy JctgfSj=ـ7:. i['vYN0XU|{O> ãӟhXɶѣ;i-D4ب{9AE2JeX2ϩڴBZ`K<$@)dyv6ۤ2P]j{elF}OLecZ[dST s~:题{OI2]br'\2Eǘ]CN IBPhY]jѪG"Gң7_{_w&m5oi@ܕKiv ^^%?> 5e <1T wD15=9T `6KSTV3M!!`2c893ڂk Ø8,x-k2YO#[ ood7W2P٬HlFmg6,Jsp2lRy{vLEAiU%l:]Cs[soj!QSZW|2Nx6|«R;KHCK%bTDL&"7T)tlYSLxͅ ; zMîIre^u©nM& QN^+ +KBzEnrY(a[6yוaݞmr[R'W1.S9P5*_ XRH4^)#2Ih4jj 'qFFet1{OKAeL̢TJ}o<:]t)2ڑZwG( e6BX(Z,ܹTsYB^;[Mh.ovei9bo+Q5b E?9F/aS( DKFQc3RafDjYu!Yke Uo$$c "i e≛a ,# wvՓ^j>|lA0J$BI`6+Dsq$}0TRL $ i-f85|lMivi'a;pi% k_Fwcr9.Iitd%eRs6jKoؖutKWxV˸2^[xZ' apbŋL RRa^K\'*; Wz @m`_ZaߧZaoZuYcríM(|μ&4$jR2.tAX"Re30;Zh`YcL{V\`\_ZBtB^V>8eޭ_Z!7pr G{~l0ЇrLy_;䇡z )(‰t64xG'װZLEy g. ą=X~Cgx!]F寅ZŃ]ksѴ77ȇ[4ܙU(uۯ\ưvrߊ2/V/Uf&5i~s~F/27Yt%'[Ӳ"p}?B>dD`[Sʛ5,}%=OIr_vy39sY7]W?رS)OR2#^ bmvFfJmpgAʌn [=-w7KH^:O׎lIˏ781iy0Yo`wιgKog{w|d橻59yɭOm>hO"?1͕vVcf&ezr"Fz@ZrEd蔓 td.:F߆R:]ət u$]R|IAJβ$tLBрLIaR̖3K- ,/ٷ|NC~}KY>'sM?l7 DL.|@Oi ~V# HgӁ#>( #6+Iu!taMӶitI,5x "fM˜ l`ZFiJiIuvVyq97Tj6c2}mpͯ؆H}QM1_ V*FEs3G sCCpQ S:gM,YZee2{U2u%$+3f&`xIi͹ J8tպރYǯr = 6 s AMw/iic[5Q!Lĝ7qlXkloch]y>93SɓY4Ak7#a|X/hm ɱ#X KwqC&?H8o+A"ݑߑȅ!!'B3)U]8nMeN>=sZQ{s> ãK؞hɶ!!-d]'F@ a:(JQ@.KQ*.xNжҴ[iK ̳&Y敱j# ? *EK]/ iJki-M{"e=T)lazwNe}pTVK,XKF` \s)2Ih 4 KM᪊u >9սg+nKjҸa01ϔ iEĐu޿??|㫣w3'~W| v[tH$;זn"l^We fwJkUX>̅QZ VIȨ» OOzmc=ؿ".h:s@{;]]scdաth?Cy_r}ttO_+|ٻ_-`<}Zf d\fL&4jeɪz+5Zƈ!dkH88R! gן~Ǚɪsˮ#_˔y:L!Q̰hD` }63rڛra/D ][oc7+F.0xk Nf`2i4x%Gq߷x$ٲ#en[&yb}ů!%)q99Îi^ Y@;(ph8oۮ2  U)8|9M{~~E̍~ }Xx%/$Bv o\Z?M pфQ8\i[ƯXS~ 0[ hBoXi/-91Dljjpe3a,-8ސg-uoxk{SXeVEeVVjeVVjeVV7'ƛLTVj*+R++R++ROK-ZYZYZYZYZYZYZk9쪹[CvƩ1 R1k5]& YKpƩCqy(P8:'SCt}=}ZK~<;_.\<%il?z{0 0>!9LGmY,' FYW6O_=̟ѧhzbo<=Nr=NF8SQB9aWQܟ8^xf4M3{6$Z@J8j=!1(s=Q:nIn3,Gq0&Ow 1ؓAξۓ+ۚ*~7"Nxpi$Uto =Q B+5_TNѬɴy3?p?O~j|=}a9[1~/A/+.P\OZׇ ;[~|tIH Vt5 FNW(*Q C [+W1Av1{ɚӋVQ/5j׻rnq>s   GmW:3zJOӳGV,wju*Q&gYo0}|{W(?ᇟ~?3eO\IJ1NSn;L퇦8|haeɷ-6㚒1 jq?zcO:o? /2 =ɇN{&^v06e󉏟uo˳EZуxn;ǫ#1b'1aFv]$oI 2-ý SJ&29G{}7Ks˟`bRFh w'\d.`@o\Z|U";ގIww 2ަ @u$,6yol.<}P a[́ c7ufi!rC3]KDs*wI,8 Q5"VDx#% dKGͩ}"SDpRc_YUTrQ[w56nl%E ' Ȇi5:֐;S](x@1t^,PJ(UqEBWHi&`6YhPƚq.@.WU6 LE؉O9T:Tl:cMV蘂NS+}fy!i'kʥq/^]34_ V@BϓެԂB<~z<:4Vv]ntu{vџ%1xb he>Q+,pCQ§D;MMAfHz뜡DB0Mnp|rİ vG#hh4xh@SoZ%4#ZFw2(P  ~BIKJ8ŴlDSpDH4ǟ?h>CQ 5z8N$Y;%[YE""}ՎyPɳ]ݩ޻&#h-~}4  ٳ =ep<[Ԁ@ˈ*\JbК'ʢ!9QDL*neNRBrhaq!|%"D*$X8$R*&4R-cl뽔b\[m!pnF&9>lx]Ƿc?rO4~Ɠ7n^ l 4*FzeBM\Вg<1$2p(G*h !?Z8fK% *!#$&D²:R)5 a\.&ϵڢjWvmrPɧ P-Q"W0[#)$%F[QΓAW&T=pt( <*50y"^s(b]H":*k bl[P?gx,|<"QVX-b ce4%D*@pjT,e*#eU}DԾUꫦUM;zM{=蝆}oǓxٚQQy5=mc[*@D2))ϡYx\Z&K( -S>5p4㶑AD*k^|0+@=I'uR%$CbwMhk/1{nE B7,9;n7nM/uXO7; xz em mw].{?ɵSؖnlvw{ǧ8 SӉuϭir:k~ٞ[mز3]zol%qE -h8\{9yߊ@=go8x9v~Nz$w9ym: 96/Zq~W|~SUl~ lnro%v8H9#f y9_y/RD OL4|ϜBQS^oʢmh| PRj"ȝ "tJT@A!T nVRkKUvJ(Z9/TnwL3?opo ;go[/H:h-oN;Q 73i浣U/G/[j ͬYpvف* +}`#Y~1uWm'o&-s'8/3$ %l\Y]Yj-{^}KםV k\}zY1JŞ0>7 zغp=\f yJ*^S U-jumv;1RS1UHcnYOlœeYIN\kr=4n}بPޔ5'SBo@/Oa zOTjFO!KYp4pW):c(N)e0./7slGL |\_ͬd77(6:%Iוj =DS[s&7nyۀ%KIkـjhԄ,Zp:҂STh)2Uمt?\ DBG@¶iZԉ529=G; Tx㩎F&--Xi#1H ;Ŧ oj-=#ig\{H4WJ+%463hfXk2;W(a\*Ku38mc' nZjny29Octj.GE QqW$>Z!!IBfm0?چHYFU,nCwc͵IN-_3ᔙ2)-|iZ's9-޼i~˥ZN9 ilՌsfUU̵PRXu*ƪXH-tqX̘*fh̪5G(:URk!0c5;pMFs$J|uJe-Z(k-ݠlj%lsL*!M"#;X1U0 +&|M8SDcUir щaspqoA;pͨ3W,ou9;=\e %&-*cŨ|2EOꉵIP7!Ī!Ruxl*cy 䬾C׍:־ܻ),Zkz` U -"ҮRYK IE좔\ Ȟe5,7\NLI:Ų '2S ^Hgx UgX!HTh5"fB A`&Yxh!ݑ 6tyE0EqEx&H+/ǂuZP) Vو T/Pt@[Xk֩搦 mgdWNCNJTn \&܆s(2! s@jcmU&Iq,X,^NZ *uJԌ9&)˲ DBZ;,Miڜ59Y{.P b+w%CA`K^ɀ TA4,~E;;eㄺXNUgdA d*WXhJ= llBAˣՠ)2X+#8١2Bi -EL\:x,,"˥trZE/PBc@ȳBoTzr%Ōq2M!̣V;@C>eBBd! ̄Ѣt,>\"·(, $hX#faaK-&lJ588)$ԙVNec\"V 3I_~N! .b8f -M=n4XUb#K` +lo y" RDbk-]ZJd5{RQP1lLbgmV UQc%GkiT"j$,S e09!j2a+֢Ǜ \uXgLD҄q!4Hvnv#e[N%TF`ҹdZ`-As#۸գNyWvi|qR54g B][ q Ihd*"mtnLv XmНa Isy.j 0灣e`09Aq=A(#6>ڣ|IeW m夋ҵg.:TP<*LGti'PO JHvEoڮ̰n$!>WĊ"D W}\' W̨SXNOzل,ϘV1 /T %x EifF|R, =AT2XT .C1#dn҆E n @9 D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$KI2}H 錺G$5ސ@ֱoB =@I"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""%`µO$Z$BB*͓@ ]@F D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$C}ri=ߵV|.5m{_^_/__PZ p->zd '9.uh9s2 N9?g0gu0;)fvvRy^ AihZzU?u3ʝv1wf8_vQ! DmnMpX}8ImQ3fSkKc׫e}5c?>h^€vSv:u!wX/oΏQ7{u{P,NwKn¤ /c%e>[S]EtJyU>]4yI{@#M-&&p;&H`:|~veeJ}>]-fǑ aR:R5gu/kNy $M_]juک|;Q9}<2!ɼs2O}ˆTyxaG?qu~2_*ZK~3bqe,*wR[`+DƅѱTߜpI}gOB@w*M1U¿׌mE0O\Ft#XnaKz8 t!g8L^uwb铃Û :@>c< g/yH@{v,/smZlxv:Wc %_Y_n W~ q=PZ>Z.bR'ǟOQ~Zw|}Fop8˗Rƻe 8{IhOVv<۳zw du ~}LJpa,{~m5{(-&>jwa{al*O6vq|mv ޴٫brK=DkcWvN=C}YΗwh>% GۏV%ATnKv%@fg' $8&6j+[Z翟buZebO03&`n9A^ %2;>RT-i(dQ J!Ji ih#r?8%!$sZdҷ &95Yً_j$wj:7.љ v3؜5ܸ?ڵFwk4Ĭ{p(%̒/즳3u߹O tCe;`avy>\c֝ 3"0 ]^]ZTz+Dz]BjI v0+̴qɦF/E5twayKrk!m&'wsi/Xoq@RrIӥHM"?v^Oz-SYfX"Q,;4+}9ޛ`4?6&ylYUne-Դڭ;:6~?mzLY)TMTUwǁy2l9nxJKu":ס諙7i.PT8x3N2F<~ k/26,[ aEˤ(4g.z &bOEcdCN`1 k-)r Bu~엘"0ym{Kjwۛ}4vӍف%g'M?MKu{&[zݧ-q}e-{j8?Yh u|к `'EALWσWSE#U_VS n8H=Fy2@U F` 󴃪TBUɰbPӝI*ڠ"N!N4*̙9DP5}j6qvCUŊ5Ä́TTE*xjI'LP\<dbtMA_]Hh4Ed[Q8kQpe//^jyn5`HtVIk0W.´C%a$BY 8ȁUX-zg՘IDk1hFxV28{hʂ}`YNϣa 5yf\ywjtG}j-Q+pG4sN;mjw@pYz8}6ٻtgf[EI#;U]ń'jSb:c~ ɭ"x.wmz_hĬ祖a<notw>.qүgʉ#YzʭUs;_MmߜCo{~~^U秭:n~&n.:;7 !oOfjo/됢;Y)yPGG(Oװ Mݞ1qV{tHӵ/ NVX۴e$ic*7iC#Ϸa!_{yzaZ|6 K >Wo$׾~w}qo觟ͿʻrT'K{#zKI'ݗ q c>*1byJ/a©`}+|$Keۆ1,L뵟7dv`k, ,N a /6.dtx6MMs4L\2afL; c3+bRB,f>89nz9@f:&lZ:_mST{z ׼YWR $$@ ` _8%V!c4Ry7ӘG"h@QAsglr^R*:d2[p6qv[FuIv|S[6 ޭ}u6 6sA;噘FS]/T,Fj sœ|[JctB ZxaOS (dLe-/[rjXTDY:ȐLwZtRQlNoIJ<dNM-;nN]_4xBD9 ni J`o0LI$I# B*C#"%!ЕRk ǢAu*YMc[:zJKuv ,,_QjWp~3qvJcFC A`3Ńt-oK zrR7;B*oq>0qY !#G b4{pXx P0vY4a#A8;HysQ 9@*fv٣T՚C l99X RT+eĝ;1?K1ۙ;߹8vLGCGonJbP9@u$n¤kr #@RGTPJm|,o6Bd>ԑUEֻ?ဣp8!9I`|qK zvHLHbV>?-f39 *X0B`D$|4P C㐲4R*O8@ /VQK #m31PKn1"T̸74;* 7(O^@O:J,PA$C"0WeE4AMhwdU XZcז=1{_Cl-zvvDz^xw>Ԓ0E##MVb @;6(+;? gzexcuw72+if fwI|Q!9B ,)+O#pK/6=Zq KHY@ZoFM޴y 3·u'>=Ln|^ga/f/s;Y3(E+|ԉ@oM9A(ePHom:@ueU ЇrX(4i%DW1ɖJQg/6jۻR{>jbaf!i`إ*^" M^,uu*@'5_N\t wwߟ>}?~˧~'Xu30 .۵@z| ?Иƛ Mgh g\[`>|1nMhnD/~T+6?G)UIN5szvHA=?fi5]R}]*&E{_0.ƵdnG|clHXuY Y-bc>g{38W7^N?Gf hexPQC{;_ZcS0 O_7iͭ8:2Io)fd. , !'hݩ#3]%FOPSI)!-Cȧ{L0P|hKO"v\+I e շuzmh$NFОS0:ArfԅW;f`tJڀd\ Ďp&$ `^3gfȻ]Ӣ Q f-@!tw-HةΑS#a';G"jEJwk6H'-aZԧLB#$p /cX8M(rnS rbrX+/rh̝Mh ]RN:; Z/*nR|߾cGg`yϙ0ms9rso kUy"O SQ!4KB"bEN(o:U!ߕ/gEmhWq2_5[ط;srz32HUᑲ e)匵30{-#chn~z3m쭳 , 8%_NrWJ~%oIE18"9E<0,%[ Fj,z`Hj䈊|0ar AH֓^=38M4܌kK(}/8=_J|>_PoϿ(d\olS]f5ֵ]{َҎvq 8S.dJ mJшIE%IʬQ(" ,GjS\-93TnϘMݞq;J9/̦B%J6yHߖ ӇuAJsΛ'ڠ0iЯ&crmR)b6ʼʸ@ Yb0F0KQ ˻u΀4$g<` %mR!ն^F{HHLBDni2mgǶٻ榒WTPcϥ*>@vZT`|H5WP%$TsnuC2Fh.=<=CAbcWQ[2صZM%YsjkIDQ9e: (hJTqIRQlq5D*vGKVQ$(Η<.֜-<^=/uH{fɮ jqŃA]ނRpe.GxVV`ӷ6G^^{= CVUT!'xx+ e޹*['2J@fVDHPV Nad$m "^c#>XαKiKJR*%2DŽ tF&gXlK|y__t Sөϟluqᨳ(2G[it*< :# 9i5%lO>"0|~i$ѯz&8e8(3!Oy%=G ǣ7:3PzA9rLxoowArɸ,'~eV^<䧉#0h\RV9\i%{,sJL" (DX$P`{ pc=lTk"j?&  `WDx '[1uh&^=]߸8v:ܞ]+B;vHl!Z1b>BK✣iD~RjYq6t";pj=iOn6ܬvA5  gqmY=\ JNRT=ۏ \:$ԒWr2dQnX 3L RJ꼯!<O,Y 匷Nj9qBħ^XDR%$ڢ; kLHhM<(ÁDVHu mm"ilGjY_. ݜdոz32_Wy(w zF);l0mԒK)gWVr\J' G >P1@C| E8HIap s J\9xƺ&yV&P 2uWٸ҆,@O RA;ؔop\Yj{u;]v5eԂAcrѳAN6' r_PUƃoZ{iѠPOBU'Z< *DdK&NYo/ȧFZG>uyDbe!SOݜ6}e'W')9|[ʀ!.$Qq4-wu^>?ǝ-9-Fxi0;.o< ?uēm.T+ |K/ķ%^6>cM>v5ĉ!YJOD#UBS9 W̩M5]%ia|0۴`xN:پ9m (Kͩ6b_ͻP7~0]1N~)Y(i知&ʏG^C y%cTQhaGq,Ya0yZ ]ǯek_JHtm`mS_}2'u3"܀ tHEAMʯei| jíըyKgkKneR "nSV!DDWH/Dƀ^7 :l,'k8mja򰦼^,] ek͙\>%A r鹫~>oQB)j򓛉 x(DYl*J˄1AP&}Kߗ@ounc3-HktkN.yN&"Q(hNI丐.8IDO{h":[62sP)c$pP(8N7Rv(6(ʰy #)kΖ.yy~0N3Q/o9$X\fnbuz1y8qK^ ܌="&SC; f S-sAy]0]1aD,o1\JJ H}=`KXe*crB9jbA+=i^ώz͓_O/&<[$gKB̥|ft},7s.lD뿺Ÿ}h{YU&+kUհH"VV8U^R+߻ z4{l"#.N;U} Y|9"N5$W,py0Q>HهΠ=qrf]oMa~p:^<{ˋ/x2O޽):?ƍ'ˊ&pBnkޣ7٫TXߴjae^OMdWԻD}ļI[Jrֵ6>a&9[&S =c$O]lm)} AM[u4(cȠQGOwaWnCW)}fᜠ)߈4D0p"RɄAHBK$Кuw?2EV?qGn9~kd$J8.Z*? ]vAĉ@R8Dy m{o8?M(&h ۴mB6xPT 㷟{>ΓmT%+Ň odŴX*gHāi > jzlXE$PH`BWHiv6rdA."ё@ѽ&.e 빱1$90M*`Cs\⩣FksN_"~8uqܹfDNFBRƁ` ZRy$8Mf@a+rDv$wAo!Z橏Pok\ l"# b+"ˆ([DlqmYRt2)\LL y2hF|hFdρׂդTnks6O_[yT퉫`Π;b!!mQq}8!g; rwOqnA[|>O7- #lp{NEFrpjI_/OYw(%ӮqtHk|Գv=b[!AaTmieD}gi^d pt+ ML0Z?3o:KeX_ż#ĉΧ?8\`ń4m$`+vl݌mXZFa{M҂e"?c%ɩO$g9:Zc=7Ew =] 17iu#,\t3J]ՇlHw>6|;sACZfB5e])&zѤԪpƱ1+XQ  Z@`7**km1 0 Z€PB䲌J:RHe V6#O O Pxd{Vh!6ӢJ猹b??_WDd%c xI# Z:RX#ՆsXWPi꼫|_Oq؎S,«v^eC̹'dOޓ<'yЪJVP}{9a].v6﷪Z*Mޞ36׋ٽ_wQen.Mk}a6}).U#>fz*Z- +Ȧwv륫egst;#>db`'[r| (զInwgt,Szlw[4W^1/u;ÅWgVCNT={6p L4>dr~zځox %qujP{쫇ڷp,eM~hMp 1(& H*J&)G SWp+q&HWZ[ReS8ƨG&Qe"r2y:e 9b9'#W̽*:#8PځV vZ9&Whju$c+YAzyݢl;禎sOz u,N^JUq4[T+ȍ4;N(ktTO QT=URj#!˘}̃LX4dˢ m%eusq!rAZ"Jyΐ5 EbYxoG̵߷VaJݬ!+ɎZ՞Y<{ٝy3LrN>zy! ~]AK7i%P P^7;!LDP˞^JS%YTwMh <2m,\(]NFdp F-DUIɥGnS18-H\d5Y, ,OՆsd.-+ d|uqnZO~m^ϖOmG|#:<庈mN}ܙ]wVǓyۧ\҃;)2f : girxuh]1䝫(ڮvoyи e+-Cɞx zv˜{# +^ܺ]4_6=\h!}{ ѱ,~f ]zBeup4F) 4dXÃ(oHbc& <5l YJr)ϒD^BĽsG&' ڨF:MLy|fWZ<6f[D[ͧOipRalݧp\?>`N# ~?9їn#[ 9Enx7̥rrZȎqXedmiq :)$q Dq d5gR&Fb:| Y&ȖƑQߊbr"!D$. 0 j9%bno=]CΖjrrjQ9gĬՋWExg"iܧ(3IXD Ih#›Ȩ ##1K.zR(M֋ht93)j\Rp#~r,&b!TBc.οw;~-.89 l/4n8 o?3#w" 3HWH: 1beJ3d>f,Kd9zN^X ,. B/() 0ؒdr,&:YXYJsxWkAjq*jʨz>'V^*IcF9YQ嘳$HX5 R&W3 %E a E4G(UddՆs[~s b58ueD="OQ/cJ9AIk@`& UL&a:N{ζ^BTKb R28˓6BL9sYqhd')qm1;jIk9_2GczvZv˶z S Q=!ǩ HSPg7>2EM 'k7Y9p56k(BN1x}+́ Ʊ1+Q  Z@`7**km˘n r e$h@e(cr7 tj@$2qLef+$6#OL d:MkIOdzs 6]ᬀVh=b>hsToϷѠf9Y/x wV$r |c\4{䆧XOT`:Yε$a*8t6}@7FtT91ՆsXk4yG8õ_}}{}fn,/Yq7} gC&;]N gQ5&X׀A}AYdJ8Im#6TiAq^FkX>O :odi zOgշ[QѬ!Ѣ4SS%& Z1bgFXAԦPωKet=",o聡] z0 K쳍YD0LVj /xL]koG+D.pGGˀ:ٛIb`!S"e[C!)qHQrmq==UݧjN!Y@ w1PKpHc@#1A)*'gRXzrT@ dZuΦW &ei?ߢ,w Fg(D 榆Zg3ԯ|I(ti ʕ|BQ"reY*I ^J|>lU T-PH*@%\ԉzp$/!8O <^*4]pܐnq.@N%eL*$ <#:j4"ĜSN#:f;tC1Y :xTpp&-\b>)U,1K? OhR0TeQ'+,c*e"(/^ZpcbZyThD*C eIb$--0#ḋN᥊(΃h18o3 RHF%S69k58=W^v6-ig|X'㛫Q R8 ٜ8O%nHL&5bL }f<;f7t=;zt5itanLƻ3ɵSUW6)>iOϭմu Tkm.=/\/`{5SQŬFI@KB@K0,4ФuVԩ}kM'6DB-MEӱ;c$:'-TȬӮcl_mQu6Z&oYΆhן!{h8/SgZrcAGτ m;2" ]j`? -2cݡ.;ww_/ZTj}Xd\=DʰxmtzTfw73 >aoO^o۸.߽h{gb8=Vʤ9qdXTtn#@EZRdAKK8{4x}0k(Gn=גυ =w)n>8>7d1qĽ49/r⟿^w̯4ާ]'yhzs2{̔ 'T9SQB9a'W "N?u/椞6xy]%P7ZOHVN(OOы\?lf9ICMM8!vE\ěޕm*ʾ-J H۱M#pC7&U,nlƳ8>R܀[Y޴yv3⯍+7_Og|XgL]K?Lte]%G IK;w@HH.ۆaH"ZY_P C= W0b>G|1ɚQ76j۽rb:s   WuWezB{黐{XԬTL6?^9~>|w_ޟ_?}/p$J2NSn;L퇦q\6C.zیkqPq;ncdm󏽟oo:oeۚ?9u^j~9-=Я{~E3ߝ(z"ehi0/ܔrxL9"m6*x\D2|e0URKFi[ ͋K~j C:$xbD h K~lG/۬#4@C *8N!"Lh hdFB\"b?2ߡF8FI*QZ*埄L].t@R8$Q/fP:2ēآ..{RvS_Fyxk֝QXqu;@3W3nƋc67%оyyў&`{Ly]lIUH]bqf+I Sv@I*?c'HĀQJ&,eZf@'JќQ ؒI"+ta<|9a|0pYDhSFh w'\d./1Q$ Kk/ۉ1q iw-9CM1UoyVP ϓ#@=$RbRBjZHM KEG iAFԴRBjZHM iAZ{@ZQHM i!5-ԴRBjZHM i!5-ԴBjZHM i!5-ԴBjZX i!5-_HM i!5-ԴRBjZHM i!5-ԴRBjZ̺BjZHM i!5-ԴD\mqe)%5VYX9 +ga,rV`Hʴ;5sH j9'nz"P%Z3+"s0" %s> WA2;ܯUfJR=bRMPKt޿da+;u CqBdY1yY*O'=q` @phG1QJ(UqEBWHi9 Ѡ  tkBPƹIB̪lP'I؉O9?LIExt6Bt9ZoiW5JS;YC^z8#g"4,癠-'7ӿ*T~د%>_}f6J*e+/qwVj.R[WJiPJ=V^  _ͩm2S4519[ %QMnp!|o-x?cEX6W/,oj˝K^i#(G!q (!px+!p i2,'RJ #i4+D<b\pm+G%Ï"Y`fX`ٗy#Y,Er[N2-#Krlʖ-GKQsbp중AݎtVsxֳ+t;ez*NAN7I㝪tuu̿ԎǷ_M^oors@MͯcܚEmf)ruhY4UEhUPb1Y\6fmr芌q ]ImK>`Ac6u̐B%MZ;H, 502g)ς,X,<*~kQOmue٤Z4-ΗW[~D6!\(cae+T+(LQQU*Eݼ0醦Alii}(^0CؒXbTVZuڼ՜Ès&.OYCQuFm7P{0ؽ0E TRt{1s,6cʂ\4yq]e襒iATt5a*T q-.Jؑ*"kH3vka<\|SAnPDD7gYi]"((2VeŪ!" #0a>W[T.@Vmght`A1XtpФ}N7" -yEp m#3$|3D Ƴn<V|Qy[sy22]\.M)Jfw7W7;mnД>]_e)Wߟn0W\c6;hr5߰& oJ`s.c=rȔg)1&50EUg"RWM21ݚsxɳ^A\w%vŎd;|;[S7v늡,IYPQ+و ϧkj6Vh_.{TYUZ)' 0VyUe;~ZlV/=}py&jm01\`l4#RT-Im&#m&S7x`io^|8_m*^m;k_Ow}) /|vsxDg 褝UBE̚G(XU)@& bsOq Qe OT\NaqŶ029  _`*9x봷#[glLdM8 wvmzy)L?+)8AYݗvEXͿyi~~tH ;if¼VpRl:*y˽Sj}W ^}Rkss:&rS˫H爬&Y#Nb_}&|5I[̤>BkRr<~Bc(/*dUTuP410FYDW|&inr<~;&L$?~*9A/Der-&;P2C!ʇ6.ĵDK&zJ/CV 3X+e6SI9G PaEkˤmҼox4֜{tĞ|r~ubqmpeKril{C4{,|%ȽM[j.R#B>7i 7O}JO珽7p>ط׫׽?p ۚo-EW;Oqs>l7o~m> kw}3<ݦm٬{at_|?C?m8mhܪis+ G7(DgEybq_5~r$58@Q^⭰\aluG'f( a, >[t6s=/S@/|,ur΃ŬjHtybSV j4FVC?Ow^RJ,2Óp#/!{N9%uQ UȻc$, 2z!U2:˽绶.o+J7zJ}jrl&Oz}rK|-7on~[:ysۜhף_?/wٯSߡPoϝ?ЗYgoA]#x1Hw*t=x4iGT/ cI fsD?D;AR+%3)!ܾC~Q[-#ؠZ"HPUkK6J} @ EkUUps^Aٝ7O>@NtYɹg:Hn${癨xd tvǢ`?3M<5g'L']M4iv@)QY27%Q+@f_Ku΅D2%-zZ`uVP!}^FGiNwT)T2H" 9N[szvbBK?;p7nLp7'ϡ]_ʇ#a*lu>@QΕ-kS *V,&[ƬMc`"YK2 F l(I!J^{vX@72vkadUaa7x(Bg, էƌwK[_k\͖;]`Jd<1b焮8HȀvCQ9PeB0-7g5bZ$WIDF klZ/RX0CؒXbTVZuڍ՜Ès&1OYCQuFm7P{0ؽ]0 .,cb^ Ak`-M^h\W3"YkPr{T,#B_6H~0.76.,B?%Il[=HJ#QP1=8WU9Uu:%* |P)1Z!K>F{7ef ؼ&-6A~C3bHP mk{"n5W7|^6v$q9 &IQc}!i0M0 (KI(p$T !93Zj-`U]2L,h5T"=&bZӴ֜54g}1]=@r'׻]vri=>dVfeNLφ;Ғh9w7Er/-b  ;'T)Y^SEjlsC䴳FFC4R9XV[\.i%ZcKnQ'mgi9wk8k>Ls&5 =[ nA31K! W Iâ_lIJ˙5Z[2"!@n Z RB +[ۍ3l>qyo<ٵC捕,5JЖ<`F|`68t 9U$ aR򐐙")" nag6fKy%hT}Oڎ<QL)b-5A:ihHTC +RV8\ir;؋/䂓`'Wy1 -f(XFy)rfNqi Tkc5"f/B[ . x)Tk8UU]^=8/Ƣυ 1 % 6 T MVj:!0T+"B`:eyՍ0`1@#ƀF4cR TNΤ Q娀ҘbHޟq"vt_װ>Wܢ|lN /3(k/$:4M@J>E!("S$O.Cgx1gyZ<[?rr? [gF 0"L?oK ayka!Yi}kZXAYEή? ը{]:l /aijWT^D,F ;4u@@ 7y娲QqHʪΌeWbjՖ]pkea[-~[P[=YJ00SƓ-7Z 7ݑmt˙Zn=Tr\uZ+:/2a=ޘMgbR``/X޲.\#m_>_V#W/|p³ιɗ $G%仍&Ғ|H:@2ibR {&É5ˋkk!@ퟄgM@ R.pW):c(N)e(\َq1g _+>l,p$7epsK]ԧ2McCVV^ wV=$+D)~;DtΨ,B;>_q|ƕT\YRaN-"B$)z<l 7GO9DSQŬFI !@Hb "TjZEw:k֔z.hCL$]TK3&HᆖU!2kݵ֜s&~w]R3;DyW} rƽ;&i{L űCGO Q!1 ?h$Y-@FIB*2y)BCF/ꦏeAճ~=S[hݝA^-~2Lvv p]_l;DHJcB JJ_|uNLZ\:$Ԓ+J(xXѢTbT5Z2a< oxL02ѭ "Q(D:%98ʓBl$D:ZD:υF1 pgEFE`a)eZs |gЮ.h-V|v׏ -jk^zô,qrj1ƨs"Za[qQ;s-sGFߙ! V 0!NAd3?KAg D Л=w) \eq<덾$-*/s.a:F셙3wR?)O 50pyB |ރ}daPN#+};S7ܟ<ݏSTZhe+y̢tByt77\Qu (QS*?G痑 1lo7)F GuWNKg^ %K#ʻ:7uL0wlZqKH]ARn@}-*ݛnFt>{|pjO0 'jdE(h_C,'v8ABjҤ&7tT kFX͢|D5 q8*m`żzWuOgtNRX+#Zw֦r{^4c!bWOz^}͂׬X~e_=8_>;|]ѿ?ͻGÿ}|WE54q~߬z\_ׯꪹumu5mvG\[ھq7?9xY*?9tf"+gЕ0B'lTӒ-爏^QEP? ш r4>yXre,qU̖T1JGOS6:X׷w ] C:$xbD HޅW讟/IuRzcs|~'ý S%rHBKѺ(]x.Q^|TK倠@ I@R8D{FDPO0T+a!ښL<((i3}~`Ԏ2w`t:ov;dǕԄK>q5ic4sIwhSf 9>SG̭=D4Y])G3'ӃdK/KN Jk隭\q֞3Oo}KÜ$lJSa|@ά l Ml2M {~S}TaCA8|u m;vz>@2rt4,WyWٰ\ H(v$^ ԛ~[j\j!dMBf;^KsozTblRo'̸ z}6 B2XeWi{e֭nQ=޹" P%Z3$'+K6XU'H:&e.rf*  cH$!F)ܶZNY:,7Y̶25Y)B*S8 .59̩Wg}",# ,c9v2ӋˁN^_C?h:t#M Gh%pJ E]_y@M:)J_'ʵ˻\ҚΗӒgeh73 WXs#r!^dwrWs7{vz㈥޾}3o<Eoq(fAp;.qVJ0Zf4c /*uzisl#Uat\`RxD4rg:x`͡V<.s;K Hܜ9yMźM ot443?/JjOw!B&!{@=6N?AM6\gyly$.p^BR$) aO1D<0^>KAYNB#^CrL'g6ZOxDY'26(]FdSG'׻]vKjM#|-=?;Y+na]mo9+}ٛ6|fgn[lfa6x[-YՒ,Slw'MV)VZc%a.C*gc~mP9iJ*\Nhd 䴳FFC4R9XV[\.i%ZrQW/IRWVh`s$)T0ICBf2bFXp,W{,䅫kT JS٥mGa0@RS .sy[aU񗇁o{ PJDQ@`YHR)\ %,A,BU7Fg:p<` ^'sO I.XE8 M"B , r\PY69-6t;-j޺~#}ZkhKTQSQE"Z5G6< 1 rpr%"l([U y*p5:S'3=<N ON8)!<$o X)#N2ȓƯ<[?zr? ǻg/~S**0r 1(:-8'M]rkФ'2h!mB= fv}]@1#,J<6WnƟȥ3"Z,'ع߰qϵG&}P3-PuDT.04u!z0 o$j<. ݳuKR Awcs>.}[MJ?C?Ϗ8Jz.#So_+z?m]fq(N-N[J&}ik['L5faB8I'{u`]s!MN)q[fiȗ!?lXA!OoI1[q\<$^"5T7o1b 5 A옘Vi,#@pҦ'%fP^mRD;MM.x뜡DB0*QYíA[dxYL35)_NF6ׄ\n{dt}<{k-SKwhpK4pmApniYwozt5tan l^rhӧYf =|,~C1˧6ǩŸc惰c}%Z^s>`ͮn=MGô]闦g>U)֫gn/xٿ>h|<;~~*̭ 1泗2A%]|RseEHZWDT.j[94^8eu_Cax{q'?"u#̧ڑ9EeT1Q}@Hb "TjH\kJ=4ZHPK3&H|2 *Dfvi1qv/4,:v?Ͱ֑sOlx!C,1ֲuL`+yvZ=:bDrBߞqQ mB#q'0_!R&s*F"F g}l S]tI濵Or(Dr* A3tƸZwmmrHw7Px{Q^B"{/"X};W?}7~n-'W%㯿}0.|RoΚׂj޶v p=\l;LH"V} Y] 0V LPR9q]p)TRK^Aʮ(P:#qH("B҉%k2QDD4Ee x '*!vQ8W2nyzѵ_sqČ`17rRA+b\E#WRI,&/esqnƈս]~]6󖎻>w动Ghg30V?V(\<XUQZf;O=8qr,N|OU;01z$_xV02[-S.HT6JiNI丐.8IHmDJIfٲ4t WQcN8(8Nj"ʀ CKS0$@al^L \O'1\fI¥> zs"cfw ^ņ|i{ӳ-eE{]0za:hr"hQo}4E%u$jB B|A9[1PM\22Of/΀F+JΈhq"DE vv|=v> a?:ڄPS&.9 T$ʩ,LZ9hsq31eA*Vw"ψ&@bpMR]}B>ݍ'r'm)^ao/Ki8+7 3k?:]fWjS욊'Y''B'l{ΨF:SQB9agWYܟ8^xf,L\%P7ZOH,H''JR0z.nMGq8 Qi>:lg{x[}w+[fpd)&*%K#ʻcg~YL0?u׊ \B{(U O'jMonFv}zEEpF~/a/#.8PцL v8AIkO t kF͢|ipT(iDFl}/ ' ^gk۳r{8ou k41ZwFӽV,7j65*Q'1m?^~?_?~L?W?Lϡ-"pB,ߟ&~5ef]s t-K>u3kJ>rKk&}?_Y ޯ7>tJ`Ի?u9M%v^JoZfy7MR*(z"eMF (VE#-cE$"W1[S(d$lg{38G7維L"hmEx #A.2Vw_2]e6#4@C *8N!"Ld4{R# !J.|BGY語N#xD{%rTA'!ĊNT|Eb$ړD0%ۆ* uR%^%p%u:hͨšFa!yreyy`{% a 3ϐwqH ՠDrN4z"P%Z3+a`$Usُ}tID'ᾁRcn/,TRAgTZJ56%^W0 odŴXRWeI80z c9؎b8XE$PH>'&OrDJsȑiHw]BtN1je6Kx_\K[rϔ|"A>PWmvPܜ9e%ʲF'MJ %ZWhBq/uy 5ͩ%tj%Ky`H:RHh) :Oj#rܱq%s%nyٲ7]Yv}JP8ؗ/'S"-w.y2(G!$-L8\I#OӋ= /[6=K%(r.oQ#2e8Q|(:Y&Z3,fl RvkX,b3"v-ƜD!5P:;㳇M`FIS`[ Y\m}f6 yR뀹q1e $K9&.fI+pdS!6CW.·:fZ]tvM_TaBֲ;9i!`%FFG2HCcʹc_{C=< =xku~Q>-ȝI1 .ԶtX8<׽[z?[au{8-ҜԜ!k,;_rme24czxA=hW5h4ރ{x>J4ރ{xA=h4ރ{x=h4ރ{xA=h[4ރ{x4ރƻ]A=h4ރ{~ʫ0}`*I~q˧t8}r[7䈡!& X>f0\3Acsk,pDQ' Ȱ1d<}IBHҙlh H݁Uւ CP 1 JBP. M#qu3qvt<_q_7$Ǝr;;?\ c/vXS~lZ&5=3y+d!OAϩ 20;̈́#BH&/,w}9eRl RPbрcXd$ؼ)eȓh}L3^uLX7g ฾b}_2S,8%ա ib0IZafp)N dY1I,U륒ItT:qPe:mC7$~ǖƕ.ŕ}`R%-8t9U rPLL^I%:NF`ـNdoI^:J=#h;y& Skfĝ|GaUo\Ǡ .U D)0eC aH'u߁1z8jLnP&o}0}>UkU&ޖ{-^EuKW 6| O2ޝNiv1;ZSko5-ZSNd"ɯe+ѿ]ʷrՉ^OKnܳwK !{]Q0#te:}`8}~^$/wnwŽŦ'UY Q<":=Ec=S3֯3I_Q2˱~Cabx1;;[,.VG:$u+]ڭ a%_uJ\\l~.gعo$ AVǴx~iR6+0*,տ2>GPqEnnhEn;7ηytTV ,*:5ZmW-:6yɻ.^eJ['uÍ4yuE֝[7HtŮAwWd5r륔N*JnԏwQP2uȤ]$(Y2&/w6 (mhI=?2oqI58V qL!r([mqH)D 1tZJxh\W50a^.>Kfo_8?ZC\xVCjW!^jM[O ,/B{>'2MUAy:_T6w pI(qifSuvzP͍4H`)O2)b)T0&/ x] d Тfvݶ1Rim0% 2&:r!HK@$%2xFbtnl3qCU- J GTdd>=O^ll]v8\b>Ĕ}ruُ-vB Q^!(!hQ)G5W txeG /tؑ\E3x2""`.T*VAw`)ְ5NLRrcNcE*9GRSHxңe3qvLDʘ eb/w}VZftZw]k!7 ^ gXvbO6ݟkaٺnأֱ-@=hYjaW;_.ڦ)mW[sWI}so_5׏ngFzy}_0͟kmk/m|6)8!~{_ DJ`>[83QoPE D"ӁQ1N_m߲_*>䧬d^B2,O%i2g6 AC\xNI 5:>ӘL }5rC9)yV>ɬˇ2=vuQkB. ''̓}ݻث_~ξLGƫx=,x]cѫlkxc GĬׇ[8_+ 27+bwLP`P/)i#} /9E 1hKfjRN uxmE֋x %"#^$ fD(B$şR ![%Hm&JczmވڽvW0>꺱ުEbۑ3!-&F>U%U18-@KrKV [= 3D<STaWw.1Gc4yRBrPZXj%GUC!LR:ce{mC{%+w{zu&1t%*%c)(X&$Vb}@H) aIlZX9#a^d5c\5yK]fzS9^`|\H1/ZrhC&#W0MqZN\sRp@=%G-";kaJ:LH"gZl!,L  5vۢfqvHr zvλv}bna`2 "/OO\._-űgBG>0nֺW,\0T10&5xKr()qlQGծűaExg102[#ta/`bo(Jp1G,"p/?I[3öƤW9dZ%XWU2xu)Q)vеio'X"4~ {9l!az a$[Q:+'3#|,u?w)n|y9?.^n |(gR\C y܌K; W j{h4v#ofU0< V,mx4zsp1oWZOצRY2BzӀ>_.YCK/ .m@/%?7߼M}yo˷N1Q7_®XypI(B;v~}1k*֦k嬫oӯr B>|obg-ғ?7Kq`uN=;Ȍ|G`@<>QQ1R Q A1SM=S(a75Yd *0wHs@yNjy~jfߍǓ㧭 rS>05u[YQ2. , !O֫uOm9t^h*< ̚2|~`KXD[~ZNz7T'nN<+ރN?1 xcƤS-R`q,W1/kK9j~5wEVls#"lއ?o81NAi)'`(k[ҡX"ƠQKa' 'N/*Sgc 漑փOX?!aP1 -T*EgQMm_33Qau=vBOC C5p$xGJ֤صs%ȷ~L~s9 Pj~g33v=;3J 7aC(`+{< l?7& ZLFqt{# *$Y Jmf |jhk*J<=VP^*W(+*՝u~Vy6Y_NGf%{,3"pnU>oǗPP7H!|U)C+Ÿ M+/ 6ǘc|:]0a{}\mzm  cgmI 9Zr;AfNK(z"}@ɋ]v)URo#;/cwYXuUP~;)g] Z 0x.69TJx{&qq $0/^p\_9-#,V _-]M))QiJ0~S, dJ {fRaCn2J2YdrPUVm(]^>_J{mfp^0M^J**,0& |CM[QOs] E;aG#8QDTrȝ5Ȩa\{ p*b[)i j԰2L&2:ÈJu;vVagiݻQVcϧEDc-TpLnrݲxVa?=j/+x2^nY}c[6A ߫! 0?, o)yz;xX]5mձo\E\Uۖe㻁J0.ǣ@V>zr+rgm4DGT:m4paҟ _zRƪGgl`hc#fCigW_TnjUn1TM5A*8ʝ7\ 3f4j|sWL#,8ZIRTMq/`js= u%(|_>LB~0|jm5ߋz$#a;Je2;}8O2Qv`r˙&QkbAF{wn=u0iP*r,rB[gGKT1461vXr1VMT%[U*:Aܜv1ٻԆVhb|n.eV#7b:2zz^I%P lNH*g@`3^Rk2F#aӷ#8y-Xhm 4wFm6*%,%! jCc l8wSr 8<5)GK_`v4fL% ~ `gMI^|$M /WAJنVH{|G~;C8о9Xq6OSvwAX߯Hfx쵲2T5)W MX5yRՠ.vNcR=mh{Tǝ0mF9+0+(iOOWy5xZDcL`^h4 ,hÄ0.J52D$+]+6}Ї j<3ɼn 8y?(z[#r{Ǣ+>bOJ䄙y"eـ4 Ŝ 9TJH餍S.=Gxpyӌ31<cxǵH9CE3(PEJȀ \nK|G}l8صu:mE+WvpKlv3n c8سEx[uB ,^k]/%oBh )rϷ/`Zoŷ* [q5/i~Y!{eo, cW!ҡNfp+egnz6m j "Ja-`6KzI_uW8{=LZS[ர9 * KYΘ KT1o9}ŻIp=q<ώ=U*YLےr4x#hi[f#2Ƹ#zӹ?Lkؚ' jCYUZeoG$#J4'+\L}_7tkQ_Ata5W+WkvAmx$#%RJa 袊hd$82' JQ(ㅒމtgv!A'x8`4~hhŜE"H1!'cȵcLP}H7L/adGם+XҢVs7^L\㇧P.{^gPRzIϔC"LiI1 f l0O FL?뢂/N$m9yPUT`y0}j=M B@dKG S)["e/XM= i w*hXHGY*h%{gmR0#4'(xAc"ngis/ %bRj+Y7bޮZm3A?`9?O 0eNU!WI^U+ѬRݍHU*FN`aͭI#Ijry㝧ڋdX1B( EmP;i",j) sh`N9!RMeF9)8M(rneZGEdQ0A[HX+H][c;ݢ*p|C:FIR2#=O-gy˙O gZ/u1G$k^"yBΕe.Os~6PJၱkT9_&fw?$I_ݕ ZٛfNJ 66c0tE@1 ʕYȈ/"#0}>OH F>;2{=A40&bnB> '?OR21$HL% <,{i-AH1{e@C<f6!fS*Cjپ@XI)xQlEwe}8M':j|ܽƱGG/H-ϫ?ȯ!%qgɉX2.FJNFJ)D>'2D`zdddv!HO­_eiV^C/'he,t$U1V+E?RFbǃ3X  `Ӵȭ^~t}bܑWwa7ŠFـJXfh7Mc#BZ;_բ,916\2Z"5Dml PNe)A0ozle:3f֡3VrF) *ȹQG}): ÚlWme[~ZpJ &<:GyKGaD4-J8ge)0{kjQלe"_P-|խgIVBv/仂sȪrJ> MTHK%k@fXn *̐gd6ԧn^ҳyqj8GDׯKWq: ®įx=C+O2[`.m: `eR鐂G< ̲>UL(%kG,I6HP'3dy<'TզQ\ k2alȴ ؕ$uh!2a%MײDNwma*Ĝ<sH@vGѹEJxT{$dr)eUԉq%sci`,ssC;=mR`@Sx]ZmY؏WK;`k$:GuH']Èa4SìXZ'L͍|G˅]9<rQ{KvRuV:pCھ;$>LcyW!?,+uw=TL4p<}>Οh0Tʎ)H=7 >~\[--l3֋f\W;ƽeq[vW@R2oY,'Kms,A~i`-xD;_~"IjU% ζZw H j&9`i[q G? A%u4 A^EȜJw O4v봎Ω%-Ѥ 9 ALF=w1漖'nsGi7sqѨm HY2$ 4pA&Id"3'1Tc&;юa5iW۩:hè=Zwfa]{钅u 3^&v3eo|)mǯFVV 2m p)|9~B}POZ5GFg16 yDI&L?$[>fApk>J&=Vr_GU}w*'ϝK?PWClvڢ` 0mטM C'C!twX^E[,Uڼ O{ iPE/<_ '@d޿y!اdZ*atTԩ1&  Hg4Z:gDx#T%s>z/dQg_L+c0=z!HxzYYG.8m`4V'FX+#ߠ# z(ó sÄ4,\H&<8݆Q$]vdT5} dk} :nAWśI=ewZ?rx!Bۍ`}^H V>'#|8_Qtt~4ϚBT@?>Fҗ]MV*# %tF)j4Oy43Q۬lΌ-)k "[dA, zD]x; ;D3!EEf)9̉ީ tL SeTt'(yla{LF!h`V#@@%ƈJ)!H@2J $՜Id.i&3NUidGVϊzyDIJZK (a)ߟtߨx4mQŵ{T5&hw@ܼZ\ܪZy $B*Y(7`-ryd\ YFGPh#]"Q[S!0<,V]ro0i\Ϥjkjӹ_3*ta5UBe]({]p5ngw$/Il׉.gq;şhr1޸. S"fMJ*pK{R4ʔfd5(@D 8\uO#H-Cq|M&AhId2v̺,!8#fזiDܯd\}դcW*kmkYoM ŜIgFOw*YSZ[()Q̑'M0>̴| Vqq,q /:J4k0E-Pب 8b!Zb>T>6a/ڋbXM>vՈFԽF5❜:g fd2& .)GZ2;rcdK+4A&uiȝi]ֆӌ@gB H##y@ 6ClQuUҋf)YMJvՋ^4^]"0z&7DnSPJ;| )p!:⎅zzq_a5U@'8H?kv9;ÓFnLb_qyEE\Lُ_of?_e?~h]a(ufiXZMn6ûj\ _\_j˫`gΧ\p`^~>1h+=qx0rr>]M gBS[&fZѤemBSN=mx[]I

Y4{YN3 JJ n 8K,%MRD~ 9dުOVӹŽQ@6etck}9Oّ~Y$L8Nj+ !=rrf]iΆ"9_o{{L>~7oO?̢ϮSb9n0?^|v<}t^-j4֟ltZЕza6<}X|E3ܞ("Yh D01^ d%Z1E\ɷeey~ !wƾJJڊ#@'Fʻʰ;ojᜠ)"`Wq EɄAHBK>W~dUI*^Z*韄\)\B _q<(xFDOz3T3TgfM<()f캿E=^X6N/[^XHx\)n*rW+ J+@rJR(B}ҚMZc'xғh:Q5Xgt>@'JQQ ޓ'=btr*\Kz/ܣyf튿l|Ծ5{x0T1IJ#u4i;J.@2pP0Qo\Z|ϼEwЮzB^h0؜xjE]Ip|7Ϥ3j/|T"9'©޹**$TF ̊h>#+IAU Zb?rN={GdH2(Ufe8J{yD*F4DY&#ܗ8`4(,x#+5W9XE~w#@S]S_Ó]Ja\<1<+L4;`9 Ѡ  kBPƹIBU$2N4'a':j>1gJ*S黦b|t vԹ j*Xkҥ׸ߧ/.3%4Eg}\}AV@I nRBDtVj.RK?_hBMo 59ٟKy%<LT"H=x";NBD(FhcIIKeoT6Fvzhp&o3 RH&ץO'Qbҁx-dh'P;@і;Ҍh\BYH *-L8<\KJ8ilDWDH4ׂW!Z橏Ik8IEnlx|}R̨dH(|>xIcf18}_~m$TRBZZ#KI] ZDY4D4'=sȤQTQdZsD jZcRbTrKJ#clFrJ1,,FB(bfY2p?l6ljWok'޿&Gz;Lu$@BjQ^T&$- xC"I U` Yٳ!8YM.QF@ivDD02mdnŨ㡨- Q`Fko| %rO5"hBJR"kĂȍI8p< ŸxwZPqPXZQ!֠%L{c̉u!@<*QY[Mg36J­= b1x("ˆ({DqmNXRt2)\LLy2F|er2u@9lr)!RQ&IEqFRqt85i[EQ@qtb"64Ge},F%EUU=.ͅVJ$}:8$3 :# 9!.wŨx ~rOzϚ3rb{onNp3y?1~ @rQupukT}ҌJQ޻:{Z;\2QEg9 aFx ɐyTR!J\>H "tЯzxp3\2?˾<HHhG/b>)$l6%$+LUYYu2ŽxًxbPvLL+r #@pҦ'Kz H-󡏤)I*q|=ioG>nO.w-ک>69ثs]^}2^w 3rx 4IUT9$a./ A#Y d}Ÿ~y3}?BKi5eZ87 d/FAA3XM7}o6.v\6CExe"nvldEeB0AHʾFgMK%G_\LsF?1֌9D+D r:W!PH1b8H7p?c_n]s놣۔wؼ-؏.b[Zmyz65НU~4+l:Ңyu:<9%l;m:@Naq`lHU=<h 6?" 5wT4-`?dr?>Zތ.2#~sHUt闣O9u"#_tV[ Sc;+ULG>IJA;R>AȻ]ޒ_mu֧?L&-=2*KC4 ڟ2b"Feڪo߼>AqMaJL|/) "s/75>^@1OҤ(t1??)e5}bz![KFES,MQoW N˼{/ՠ|c˜gsHFZgoZٌF;̕A`˓<9uVwQCY|[͏UɮAwMTZ%wNFu7ozqh^m^WQ)yg2l Z]~!jC}MB kxMnX2ю }_Bl-(LfgGUکѣY[s?3ubTte1U%FE}L$>G['1hp'ÌߝgZFٵFL.kry1rcbL8CR_?p昕ߜ~vz6lԡf1%vzA|z],Ӂgl_/-a1:p஽Tk $6{d{uJܺ8m:EkUV<$Z;<ژ[@kUڒ2ĬBk݈Ǣu]Eh6Q>v;0HaӂKrm˹_̱ʺC|OR Ąsµ k#5ݳoFnxm|C9 0^/fc~ژk;V4#1}ԯOzv޶/=m{Xl/=,s~o# k8+#/{\}jl#Kgʶ&c}v}<~p}_ $ebFR~ٛů|?q}s?i7(H{Iiۘmf)ƺOkMA]7W}O#jSş2;d߽nUnzkD+yf\K.UVKSziwI͋onrTǧ]J(C=6`_of+rrvhw]GD]uj`ps2X7] @ m7SD b4U=t#O#lpZvVƃM] ;Q0ֳfl/<`]'u,8C cKUԥ(!.Zʻsz8|p|S7gVӬ;#ެ]OYw/}xd^Jl\[I#-Y162>$bH\8Ò8u翂R6WBb|͒JKr&,Қ}Pa}`%R1'\`əeIܗNmЃBV:'i_G9CTR1!)q %QT\$!2'_w1 גpw_%B 3!47cndSȲu%6=DLZ40d Z  Ʈ6T1ԙ&IHxH&! eyYβ5\q]t+,d+cCX,S0 1w@4uESd4xҗq~2щ蒤L()*?:CT9Lk$)8RYx@mJ)%*\ե2^p`?gP Ht2 ,KӤV JՅ!rRAhuR3kik=PWuCn`@ꢄ#I`tdmT<kwOQpobP(z=n.C>Dr9JV N `:b!}0T‚D4(Q .#*P P`@H $]@VHz[A@Ez;U`}m^%êfcB/uEhN)s\*z*DxyY0% Ep !%L <PUh:@Ֆ"[@ BDL;: Ҩ,RFDn`4 jZ \,$-d= I3d*UcekVK:koǒ O،`/Yl=Ip$FW1E"Y ߷z8(7SMpbӜ险~cP&̌eeUueMF)AhS9X7AM-@!H<6.xN\ Z T zVe*QU̮pD4JGVE@aK* .׏Fӕ! WLjXJ>ƲFMOn!8駕s ^ebgn5'ۨ9Qě1 ;ܵepuGݼ$v{T˗D VHAu/*]YVDIRK$IA(@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H}$hi^ @%Cjɓ@PKnI2{! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! @2#_ $-g@P]_ $K{$R1$DHSC$@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H}$(V%G@Z~=FAm:yji,@_" d@# $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $D=NL]OÒjZMO]Bp j2ϻl8}X]'b_ChV8ٴ?(@z7~Qa8e6{ߦ;k,{7L7=Jѯ_5ۅIž;pNp͇Ni7#'0NFhIykc"?^?Ƭa&8F)2%j?wTrC%2⭠W5Eo4ϔhT|Jf&QmMf3yG`/97ѥѰ$;~{YZ=eb$6VKc<[2C5O+\N8sM6:"pnBNmmp[2m& YΪ8AH΄OO,(;:ſux]^stP{ pq֏XovGcVh~P^ۅ=O6 [n\v+l-v Ο 1!Oi#B7ٰu9~RƫQ8D]XFgi#?'ߵ<ϗy ܣ7KWi_Q,} Ι"t'`':oNjkl۾ v 纹H~E,C Dx!~pW?7? G7_:5pe$pvgyKn=W/3qiw/6[8(Uz4]ˮe{7X͵f쓕޺fIBXHkfSn8?xx.a6?VwvmצAMV^T7X(J*u#G-TpÏ'UoW ๤q#CyQA@~r"dpP7y֐ */, ( =?7;tgI展>N:+A.פ |6~g6Q#/*%FhIDܗO55h晫@MEq>䑠X}'wrg1~xK[oe^MWǧM6r󾨣>tiȾf)`w(}I0>FB4.{_7h°MLUP?9y]&5/*MLAixR,chQ?{@ֿJ5⤒{\Il-˸(^$L0Q[ie=g,s7O& kSQVnaت.D6uFڥ--{MyMzMRR||gH۰?}{`CutYG%!|*B IJZDZDR2KRF牀M \VǭY n|ۄƱuN uC;Uɣa˫v\0}M9m׻Z(䴝ֆome[hQZrBH&I{T&U=bq+zjo.j~qt:Yɡ~QW"m&\Y>L_^ "2:/B動Š=#j1jq?49-y皯;G_4E90`;Jc yk{cSimIpLُl]T?х ͜R?|qW\^! 2 R%k.H 4Hzji=#ATcA:R O`"&А1{fe 6K"Pa2JÜWJ[M;[cZuL{jnO)4$ЎsJSOG"9Yh!_Y,9e :g!2PM'6UlժYԎax<M c/}?%cy9"KVik'G,M"ɣ-OYi}S=]XaK.ϷѢߪq薅ȩ-TΖ6ċ}`5hְh\(r^hRq :~(UwlB8CYY~m[Bf\#1ē<*CX#(e7"z#˶" `Щ0Y/s|gP1K)W :XgV0ddԖ7e恹2JpHSmtJ1 ]xr'Iոmڢ/%r(>Td%+TIA"'BC 4LSFHe+[c.hD*ZGŖcc@U.i$eOx^쳍E VaĀ^1јLEPKxU@\ND'€Qk$4NΔ,lORP{jcTG{ٰb-S (pFt63**}sh]`M 7Yd0:ƌ +@D1Ϝpl@RJ,q<"z|nG@&8! V+9^K|Y2dOpV8#^A۲'ay\waPG$_%A7nuungޠSvWZr绮i5fb% iSLPߕVې\T/hyg1;rl{C5ќvrDT^X!\0?ֻkCP5H.{eZT5םxFf=ԝ8ߖkzIHiC p$7܈`<" z{{0a!a]xߌruW*=ȢCY_?t8MْObn}FW5.2X|iq&K3_I)vn[m)6a lBم[jԅ+w ';yta[WL( Jډe6o&i::ơKK׋ OWeJ~<3Ԛ.ԭ~ޟLǷdQ\#<iRD& aClI4#:TO $5Qf)Ӳ)Q  GAYbݟ#;f#5e_gxj<K,rT u̐v1! GDZό$!JʘA&$SuNT!I,}9>jbAC!3.}>sOӑg||tyNc;~fgw9nl5roIOP=C02bz;S8e9LXJ[6{&GM.XP/.qZ=#g* -;\D Ƙ+DؐeSsWiZKzLkO  XHE)1As F8P!R^;@^kT~~>;5gh5z&ť쩚֕nA,7h}qu^yI %+x"pR:%WABb= M$eD9㭓D4"^2Gspz%$x km_IvՋnxŹ?]3)5loe6j 0TQLE#SIb4&@muPWU.'D5 [.a~[:|W G{K0퍰A!C`ʊy*պ@{kW.Q1˛9A;u;*GOefg S@AQ @BP.%744uV*{#g&g7K 镸gnpI`E"kScQh/bc )Nф2"X@ZC1 ǥ%Zh"&%M3z3n3';{w-{]k8ezqcG3~F܅\:RlhNI䘐.8IDO~Hx78nF.3P l v`\q1Վ1%Ha#y@ 2m _?L|oةLIhZ˺#߄y'7[==refhؖ'k4Gnrr$E h:wBu3-1^!GH:~0=_R;fdTH,y4U%J @DD%Yqcbq?AJcsђ/n"hʇ+k|n@+3z$*5 RlXݱv;TiAQfB% 020bY@' $ +H.MS!6CdV)sg.DM8)RsPPE??xޔ~cT冷,K?;ʧ0iW.+SB wZW EV[r']!ht@zxI?WĀC; w~{8B5xqMQ_ԒDˑ7'$f/H''J7zJn(ltedՓ'kІz77Yݛxۻ*W WxѬI#AءCM8&&(h(~MH-o\}-̨Wݞw具W bU_t|Z,3;(BFx}U;Ҟ1'i$MGbHMa$̳GRG >ْ緓_hGLiԦ2tq>  dG>?z+8ްz pG凚*'~T{y_x>?=~wݷN>~_pω& $0Bv&;pgzCmkho24|e|qMSnwٸCSZ cz?^};cn'ʷ?fVMb~d&_AQ/Bq׏(2U\ EPa0\C~M;"}{}E,]!_ke0U S%yfu8y8O|vwO^5qGAk+<1"tޅW~dǧ? "ҟۭ#r yJ&d4A FBL"[# ]>y]^r F'p5%Q$hIP{<5xP\f;ЗQ{>}4vWqeЛ`)oV~Z|k#+ "B pB dp.yѭ-=nҁ >2t b@Nh]9)s2vyN _IWs'cST{e'/,MO$.&@N+N#(>O&r5Ik> l[.!O ֧,؅oZo8̿"Ŝz6o>[޾ЫKf|˞츆֣7MʌOK$Æ$!STWoeKURӰ7_ ׷+ߜ~վ:~>6+01F{d>nՍb̀Y9$Aw@2Goz9. 3Mo|ֿĿj/wg|͕-ll}{y] ؉Oٓ u}w U?ԗsϋyINMkKOo;Լ&n2zmsT#-8#0JkMr+vdhɻl&Pg2j2"pDH1|ϑG*E %NRzsg,O„5L8ux^{k# ~<$~lCfdف2쟟 }w#R4ռ@v؆jmkۼv2j|xzMs^;"Ou;u6ɝ&g"Qw`QS/oߖOlD̜_MF{f @/6r^RtPX٭8V:ǽ9smt6.@]mhrTy.#KI[͋SbLP*ˉIڽ|R?k+6Uul!VlHQ;>چ79|}1$`cꓜpRejVQ?~8 ys_q U|1zXS>~#v7 t~7Gq'rVQKqnTxuUk"IX^CDemV%sRfL3L>8ӓ ~׾(.ڲp^Է.f(7[i/jͤz^{?6DD0VS{T7\|ӻ韟&ڤNA *7Q(MJݲnχ^t:/׳r! Ug<ɍy13[dFvUSu|$ϊ3x{i}Ϊ[=:`dͨaZ걫Rj$,ts( B濷\ 2nDյFFcFƭFx-8u4ŵ%Yܐm֠k)2Mb(ֺX'Zy4'AY אSDYClb*%Z)$rUǕ.,HMEHmt Kpn0%Vtx k"V . GM&\9-Ee"":-P(% b-SɄ%vfġ0*hDr@*!/%Q\.cG,wqI4\dއXxvf1<%&MJP k9 z@ѪJ(Qkj$*J'aRO)cH19'-ljUSAJJ#LeFŀv4l\F9EB> AA\J}2ɮ3Jt"T]zDzd'[&})6@=/-V !D,AmTzKtg$d BX' H rNcig"<(f:>v#]7cEfJW5gA'E"Y Ä!6K$co0f|) ̚` ,QLkUr QtGdUmj(lDC@(;`Wf֑5#yGx)(`~Gڐ&zȮ\F)^eSoլY$E̾lD_=iDWBb|M6yAEm"Z8M46j"| $C aUhޣƻ ~,B\Ӥi 78;b֭݌bvtзbMDq|RT1yQ!BNhmY;3tፍ@ dUt4ҕjI#J2a1'8;্j,(t_8g@Z PE"LԴjPyUA>xmBV58 U8my4 O×Ud 2inuG-8n3ϊdP? "|qgE3#md-ŀb$SƪHgc .N&!z*`YeD-&PG]Jr P _@1 0 9B5t3 hcJY%h v% Aر-Px˴{Z  b()c &˽e;V Aa:(%Zk?m,̚$fj2RA ٕ`?Aj880"㬪p*y`* BȲ$@3A6 ?Ft26҉6k M[B:b8a$i,e4HX6 ئ{ů%U SaCjh\6cs^ymvatY n%n눙$KU#[!ۈU:9lѓХE0I0Nd2Y]ۆaMEѳT$$QzhƤW #2=6"}YaFI*ျ!(A/Q[tCۚbC (c= Ӯ}:UoX|aQvxxի2qԍ^3ѿ_u}n{V|j;M' x6e(vjh)` <Z}8x3QUJfd-l9J6_t \'6e-[Ev1hStHIc1 {Ҥr~b,Xkwx7:S?OIZq>%! |uz |m *ގ ی++پ^ YvC+Jum ['*{TJ ՝˟ Y9uWs鿀?n7 q+ 5x~>\S-8q~U</C/-/hB[FZO#pm&V1ex=p/N .8=NKQPTɬZp]\ѪEl/wk#yrnOetVS{WnFK.FRwk5+n Uuf]Yts*ݘ3mE_vQwϖĨ~].V7~}<:ͅ3#3uG&KTIIJmOEyeONFz'o mQDתf 96R1g3 ޠ@bLqS7q|î-%xbCwUYĚٍ F#]ؑHUUniչU,S[~ɼv/7IZ/sޏ0w}"~v:rMwV~4cjuF)$NO3Ogs9giLs&|jb-QW RQu!5ݲV)O=$ߨW9+:clܓbn'|T.Y|X12rP?A{Wߋȟ6^ulbmeleH+=I2k̡)1@f32(,)XB(Q(׽OqdPuW=k՚ 'ʺ'giF˧0bb.t?Q {o 9IQBWq_ 'T݀D*Je F$hRU fh1cgUdKM]5A.wnΚ;gDKtW8øDw /=œAO־{iRvo|uxw? s[C. %=dҀl<֢Pڦ6TPhID)f$Yn#Qs `pːtVIuȉD; 3> ={IOc؞_|.կ-a{ -: kjzlQdv);KsݏX[['z#.^:#dp>3e~DƋ`i<{ey ZR)Fw5l릧$dn .?LwG{(0V ^x|m ѡb92탓Y[kOvV)]l$bVڳlbZ.//>2ΐI"2g/A(@JTZOoa>TYkcHwQ˧>|=4W]M[I`WoVU.LZuIs%abȹ.XT)=95mvKeԗ[r;lz -^J+!6MvXѠ^~[׬UNpH-5P+E֪:ޕ+Mj-6)'t6DґzD%w@#PN@EmV;1e3bM951 Ro3["!(O=d8lC7SHHpm-T\H@k\RRfJnZ ș#01RQC!9 XZsE$!(>2Ǭr$S/%>OOfg#ϗT#N|EwmBJTJWCZ_ "KøX~|*D׹r44I_؊SbSJo:[j]'Rgcbu6ʰLnv5fr3mD:;7k]M-;e6܉ iɮ%ͷRdG7Ped{4n 19Ģ2OxiIp@I6Hd3a,GY/JUUx;$*Po5x# .pW):c(Ni˸t]i}PsƮqh8ztǸeyc+zwOQU{Nt, #MsS{[%ى$ Pkky_c2=8dv8}wKyG]*ň<9BL4G!Hp0gx ie%eLs9*嫎F& ;O` '?Xtlc!;Ԝ_̩xb,?^ӇlQnuGM,g&|(%0me⫈X4_ iKRC*s(@UTZۄ L50'͸ejOU{zTU,NԳI^B2$p4yT Asܐ1eޘHkY1) g1VW=|b}K.bTL 0KNW4vmz;ʁ)v󍩽F@8?^1R`Ve6k\F=Ӹ:My$ԧ<}w`$&/Q66f2S:uK)phּR_5msqY7}ݓ5c:/Ӫbp9M[:w?POƏ{[eplsawgՋ AmRu;۷Ɲe9kdmlHuc)xY{bmJ-F;l<롺{&,%Fes-RZrlZy=g  cH$!^Sj@[-A KH!{ES;LSQ󖁀n(+S_6[0hV]«ȏ%nijìAԎ:Y{gS)݃H?DH eGiS䮕ThnrcC eIb$XDḋG*F؏KO 2esDʐ+)P5rL>-֝-v)e??f[n6n޾rio;km ܻû. v}X[9w˥tA|WWs\}PmNJx&GY,Ϯ wj?{m/1. ޵q#"ewHd7 $8z R,lsӣ G9 hzdsW]٨7XCiu:)=JItkn>+P$t~~q9yUw1ź.sCsy={ٺlā \ma6}M'MVY 6]uݥCWGw>dcM:/& eR=:f׃KKݮƧݪɥKz=cNN>ny/Kϥ|vtty{{ }ߓw󏓁YCekMc O~:{֩P~c6nӬ׻[$.Ro`ٙӐvlئw>D폏$Ӌ_ZDܻby;E X:_]rxxtKq@"ZOQ ),  JQ: @:uO=xRqmę?u 4ԅW ;_/W-ZޓA8Jr\`"4 ֕jn(̪mٶ)^J}!kѧ,`dtN)T%XM~ ^I,SO!'_ƺÍ՜bv1+FfHR9eGQ$7 䞘^*h*݁=g}x/{cMy蚃qh7ei/v|~ao~r<iLSe yeN)+ר ,Q䙓rڒp ?jЎ( xts #yڝlLK<]"lr{z~0toݛoޕ>g1ORo1} 0|n PF1kNd%cpe5ԅX d%U6ELXڠ2ɐs-48^8Є> HZlO;ٍ,ɋOMgCq39abxI5rz\`~y4CIQ{E{|M/O+%,때!f)c5L.jAdԚd4coC]u̼fnڦ9KAtš=IX#r`EVubr_&mY8MYc+&Z%i2hUbPbBx2lsg>b, \@:֍?ע#P[&V[.[IRZ/JJLNG I$ 6I:i "UmD5o%>(c6%0?w>a  Jc<'ilQNf>W|'6_ /3ے?B|W?Taz/u}{U2.ާZ?|6=Īieqx&@uO %B=2+Rga:gtٳ|ſhEU'(oɸ(ӕq9$wHS(OD)tRNbsՊ<?]ޔWWo7[uKEvg@[oAx&w_`FI'.y8-$_R\K?}4xu3򯽻+?^:&8[10zo'rzx_ G Wq,4O!FB18#  sST^<9m`]źON?-zz=՚"U.rhԡgՊri=` _Oz.ɣ gVԯef'xrq/Ջ7ǿ~|Wo7/ͫ_'^k;O៻MwgCϛ-Uyam˧/&<0 1$.w^~}^L'X~ZO9kqP6٬(oowgN^xz C~|◴&Xˑnc/;.b Ro[qG[AKfhH1`2fI(IɻΫ /VV+Jc:kȢ$S.J6{ [!V[0Һ,,t+suk 8F\=NĕB+G0X;k,_c$;w@:;ƀ@ c :L}.S <YdǼ%cTIKS gHbSH`,EJHeB.J$Ec}mWwƻj*>pʕ"/A B,ҧ3*bʒu@ Ţu4*AQf2)ח+fYPHt'e:xχ.4[UqJV /6z[Ooo-[`JH>R0dI+`('5-_RBao豴c4 2 +d+&/8 Es evKS7ATA !P0hIGF25 a-1:E 1zH̚XAMKd =?LeAԼEHMQ^8s!$*kM +d $ergWcv#(a{@ RDZBEe印RHUIK D,2J[x 1aw<믙+ )=3:1]ǜSڭkƵa 7fMؿyz-ٷkL?_n_E9b@gI ;*tlLWٛ4#ݥ{ 7V;xz;҂Y3 ɉy,u(a='RbФ$:)$Pۘ\DAXflF+k Et D U CT!a &^Rї]Z|3^ݰImN?كU~7׹8i3I %(>{MϦAOqT&;_AO;=ǧ2FQmw ުOߎq6LI'턗$^[%tΝԜii5+*Hw^A.|%]<̏^Ca:dI!hD)p^,dESvU$t T {"wȼSA׏!Fo[ RiF R-O.Mp/;8Smu NI6 Zv8<Q (v R:OH,eDSo{7rVz 1Rlݨ}Ic}i DD%$f[("+gХt5ݵBtN)K=`v}Ŕumz8ҭ<՛} _.N89f.{Q5@?%ѡ::l]]b #SNuiN3єb  KUDJKP:xCql(f0 |>$@Pi`KwD%:Tml41(Hedg, )΂e27U yUKKP<ϣFW)e@$NiI:qΊl:+j~{٣xd\OLN Jv%SH`z[DY&HZ[ίsq>]@ן mj7H( Y˭.Re#S pgӞ"2Zt%B$0-1qxR)Jr PHͦselUf-ƶPpp9lW+2]M~ƫ>Ѕ'ӫbS->e#aT٥dH"(+PPB+ *PQx%d42Tb @Jj6e*6ђv}hy[b."XFsŦIA1j7milhGn A% duR5Td.Q&0=,z! <% ̢Ύm hUV-\v@$"BhlM~{ؓEX{5bǶ6vE\[ƖIHe N`Bx&h%fR]#mјiQ a07:3UD!i摑 P(ul:Fgⴛ+!:Iɶv5n]\'Y,>lEߡPI>$(E 0ހLHzAŏvqvPa3bc{=e ~^~$ҍJ}EAHя+G?yʁ؇ΠAcЧw*:LfSLz u:C u<Z]s`@"(A2&*X(cQ*By-/"Ff́ieI+TLZt$0de>FH ='|žkQVVW\U$.RPlGx9^( FG[cU5Lִ`テe0Ɂ:^o4/LY6+6{` ܨ@qm$ʨ5HKЀPT I4NЙm"t`13WLSm8[iCs:|'>@sN nNl傸Z= y#ۧ< bxh,'/ D.YlB1^FzQytݶH G\;ɩ]l!Tr` nUڃ*jRוj7l2܎)VGτ,BolN,id> HRjLp؀:&N1Ȕ+ I̐Pi+ 3 ־uy>oU^ƎŎVl^K>^@Gg NJL@tBk3!SW\0't a XTP#Lk䕓 C~O<цfr1^j,>Q4`,,L4h1譊*t0ߨ)L~YOi8Fia^Dh ǹԤELpL$5M&J\採c* `oۑݔ^I15QU]yEF֨>omZPcx:SɩtҤpx'yFm!B(ڔ9q;c O'٠,5.f0I\Q20KBd"栥ZV @:Du,(s`Iq)1Z x EFj 6-N|{P#{'s><ҋidqnx{f09;xJfw _YDtBX`\H̕CNXI)Jb[^te9pE/g :g2=;Y+zzɳoŢ h3ARr,k?B0$l_/yHc?%ǫz` hna 3 (}7#tpzd"c蹾_LYWU5+o׹P}jlɕp[Ud?)2#)zR3怜9w PaafZx[v#jyuw<lFP5̿tkh[䛾혷(79tY -s/7=Xe5NӒt+JI8ܼlţ7Yc9+hs)Azy)0[rFPO=^^?/J o@>o8{hDBp B!$0řρ2e圑sA"-vu^Y:;?})C sU _z)|3ZN~d+ ԏNn= le^;`A(RКd`bc!@H)&9z/4G/WVkS"hD+2מQLx'ʤ9V [כe.rkKFy>rU<_mr:s/|bY)R2WVq g'V.An޳ Umi>˗7cX<fZV׃|pݺ}wAj} w=!}1-EY=x>Ў~9kW!|!~بۏVʙ(d{̈j6 f߯Aݎ}kg'|D o&9ӾSRE:BbO^tf-"1_ᗿzo({89wfsW-?Jtt~<7Vy@85keiYx8<]5^yd I?+Mz:Sz VwLKAQʲLG$i's #X:g̱ױߤ=0?]'=io9ӽ>S \oVClF-vdb;;8!w, qjwغut`4456.k(!Lzo&0ȒY,h d^ Z8"@7**+kmqoߤYs̢:w,b|n[P'^N%N$B` 1b6 s,b 6, 9&JHkS`*(UPTFܾta&Ebr"%zLJ1H^te9pE/g :g2=;Y+zzɳoŢ h3ARrK˚O@k-G8.[KZiF8|f깞4ydDŒ~J8VʘY+VIEƾjr=_23̺Z'*Cw ^JQmd7w +K5xa.B3+U \?6O//Tn/DϷ.D/z p bx>:t=iff Mqwx'emwu8W.{}#z2}oRT߯X!,%\-i.h"y:0 KBר{6o;*9jdٿ}]ឃA\HQ8o}ɟd=S u #AoԑGJ gOa<-^5?{Fnl %hTUq:UzsaRi1H.)+p8$%-9i4N7}.Oo謾غ佪oy35JlK^`Zo/wC67q4ҼxupQO(M C}yf5tb {%[5bt{yO ӚJݡl^zW/}C:w#^57\V oC uj]*V:YYI`ȯ9u,a~i3hr5=.l ޵ټ"wnKgnl'fw-wi_GġR^aZ??FwW?cqg-}sGICr԰ԯʣtZ* ,<'ťw|}݋j3;(Cacn4de'?VMk& ~̟V3Foj{:Mw<ҪdV? g9>ϼN'/,٠~74}Hߟ%_"#e68ǖh+h; 050l}r pΆ{[~gaJ;m\0{;*ƗD FG1yUy 7AtߛԔY~-3E]h-*Vkb:I5MPS]B]9\CeR > ;*2aԘUt&&zZW{Wi޲Sc͝(uC4OEr3h@^Z3I@I6Hd3a,3ȞȂ? ,Q P6CSƅG^&2 B♒^(PlY^l^J\ 暫*O.EK~ߩR#B6 1Q~$@3<0@²Ar&Yf:'юF@F& ~9O` y* qugU^~)߯.s3S}U+gT@ʄ6H v|>ΥYc杰cZSp WJnc^`%S. n_M;*n8L>JEs)#/oOig?-gڹ6ۜ+Im~sLw2V Em+'ь"*NYgo0b=rOvHv紃;EUz`T $JAM*ZgHEH\kJ=!&ji.*쎥$с>i*DfvbٟTp2v?vͱ3%&NቋS#I|!VsGV ۓ&tyБD˃RՔ@FI✻ BY.D:ddt0n ⮇ڕCY%=#~0v̟ĸtmmtt{ZIfɍ8]ab~!I~7rێW?{7~>['n%/2L [⼚hke]\uHyJnRmUJ JJ_}vNL .eW^Jj+H%CJgc,\]IErQ$[H:dM2Q& o&9%.U/cp,S8$ڎ=/Iw?f:|r?OJČ`17rRA+b\E#WRI,&/fkc51S̠^NQYSOݪ(bd-eN='f%='Ί<»~[jX:yzXA,ɯx<]+̭O΃ l.]Rl:%98ʓBl$m#bVHXYg y.4^Em9_p;m<8v(v(\Z4CźQy W<'}_ZȘB77%lfmYΔ|Ny%Qs蝄IY/F-:uYh2=.#SJAi M! oTገ'B\@s3IGQ2Bk,P(C0h|0NΕϼOVƔ՚pnX݉f軥驖}&gq#B( 6%\?Lbܺ Nq,E# I"O&@s+ ̜-L*7&1ID]}j`P1?f/tA._O g`C6래1'pֻeN.6j4{0J('dLN>U)'v:nps2B5xyjkI䍣eR0zpMvM~W5!g~dC =[e-cmJQ‘ni5tbrNjpi$UywS׾ ?lƵ8E$@֞)ݛnFX~]yW_x}}u|fm Kkf9+"Wuכ ;\ BZҤ%[lk6#hmfUT̽7:j4#+h8tf$2Chu:|_GJEOho4 rf[e⟳/P߾˻ޝ~2}o߿{pϡ-]< ?>a)+47okاiYߧ]S[0}IJڠZO z?\zҟy3wr͜y;C/ongZ;>h+Pt"eUF0PVE$#"oc QJ=zH~?uYpPo󶭱'p "<.2^|?}{J߶Y'7F8'hm&UpBD(8AK$+D Y=GN &T+T? 9%Vt$+3%Q$,'6Tm+&ùCJR-{頇nFyb5 C( (z+iK/P'4Q5GLNSr|@]WuCx(#D"),* R z*xi"/GUhWuVgtcsPgsR)q,k/oO/h4hy8jhDv>KcE*F)f>r1NXN]ʢT[@n^AԉѧHdοyrE 11s/_%ɨIJ#u4i;'\d.!0Q$ Kk&o۠c5OS[-)@W<]oZZCkmg<}#xƋ'D>ࡺ%JQ ꝫNBekͬ!^!WܑI0;ʭorQ'ڿ) _VRAgTR3O_eaVv' BdY1ytU$Nz⶝nO(vÓ9UTI0HH1)ͩ|`dF,HC4(DGE}M\qSύ$!\6 L͔SG'"<5n_m-tLN'xө>x5Ep?-Syj$L<р|^~5U0[}~5\3+~RtD*P_vVj.Rۡo^('C/l0Q#$KF: 6BgJ{H_!e[Rl0 myHS,$^R2.ۦTʨ/@"[ZP,UIɧKSfrR$I$H\d3N&[;8{1ocG%>}=%b fu$Rx"9-RB!-Fii5gR&F:Yhf(& ȎVg˧<$@2#9@EBvœ 8 5[O3M69(jzrP]V(2M7{o'J:9=oҷ6^r~lǕJhP !**LJJ* XKV5xL,Z sQbH\ښcFdVyZl̒ lPDO ޤs>HFjFz\VӌC*cYp)Quaf uI&'vBA2K$E>`E21*lb0ʔfaA s, @rDN.Wݷ@P=! ) lJM0a%1r[Rfae*%n0ǂմPVQ[u1ح7);H!GEZQ;1gYOE3)sQLCdyA&CXC)h c*mǘUddٍSk/x,XM?ED]u"nsNȸ SvNFlr #76ёxL ꤳx@ݦrƴ]ֆD r,$I~d OZp"(jk&Q5.Nq^g5-9Me\4.vM@]k5&pQTt!BȂr $DbG1̮cjq(xhgr%Z4]kWhGo*݆Q{uя5я~ڦxCMʓҼ^iD(^|wF]Y%Fͭ5-4ȟ2d&\4h 6ѪHwa~[CTpOYS,ԓXXu6t?ѧ y[$$I@`5jXG#-\Y(u)\i5vVGHN}V>'6ጓmC^OZ߽ 7w|R,~cԷ>l,F/67(̷y{2lky qJ]X~W,+wb]Xb]X~W,+w 56a]RW,+wb]X~W,+7q'T) wnRMEcuMR)Cpnpe ׮;j>ah֠ K'UFK@.3J1O,zf_$0Y4@ɳdDrr-+NhzR EeJ u10FXa l,:PYù!H !irwZQX{Eyyػ7u )"WK!C3psCDk&D ڐQ o;29[SRH QGi$odZ?\cT'42%G'ȒJR]C:$3 ;YɁpI~ceυf6ك^q"@\g+{"i}.-)q",}`%TarE۷efmMőv)V>ONACZCx}5z>jqטuha$ޏ׿?'Ect]3+Uz{q0D+_K߲p[ `wJ)nҺк>g[&+]o}p6& b֫db`NEFλ Ɖo7n˹'zBT3WUgOO~ux7OCK/c{5|/E'5tRrgG'1t* jB, h.l%ĉ](F,gd&}xT.W5x. {YuwNrwŧ{0nJP6ׄ&s+Ap(̸1rhJVcq(~i0KFex43AK۬lΌ-F5(QdBB!nCESZ&̢.HH#:",btUgNzPUgOm+dS;|W;{ CM%kHl%}V|"J RIrAFEBfk&nCMxmÿHUӓC;ʝ\D٪l"k rFdVyZl̒ lPDO ޤs>HFjFz\VӌC*cYNӮual '70L6Oσd#v>5[MCq|(()5„Ė&oǬ9ʴAjFl ㊋y,]M;EmUU}.50Vqq,q /:dk0-AqL%_yY>İ8u;k58ueD"v"ٻnU:jZr(.ʸh:\pEwXiǡxh+= ʕ(NkMn&tٮ{$rdӮ_;󅨏Bo74lVD.%h?uc2rAi, +AcV`w#=Zql)OkZmW3tC8>5ANR82AmMXw;V* aMuOrpsU13ֻ8o{竏Vq?L=^}=cI RJ'M\{&C \h.j=k,ml 3,[&_9dR:!,0.]$!'"!}R o, h,5:<"vгz:ڐĢ fr}JO!%whXj찬 ށZlˑD-!O)t݇A{M8N0rz <M2L 0ʑwpHGd!Y}5=c;_gwݗ  Qg.J/d%TċOzx :LFc~0Ti؞&Q .{t1Y(xt6IAhm~6C]%̧t>R'~Qm_c~ (DU0Pzgda ;9V4m~NrHJs6J췈|_L>2 ByEy^c@9^zWg_{5fݛw&^7 aZ8_kWԼR%WYDqDQӆE@RSgc,HZ%t陞nLw/W'E^Q됀o鷪֬HJB'ebp&g}jd^'Mfj}4,o\׾l0&YG9WHn]/^fj$$jDK9 cs4X?$5,T⺲0-aÐ*:]]/E㦒>6`JfEqݗӺ٠r*w'y{v#7]8d6~m 9MJBWn 3b>xeZ6]p725tqņ-\,sO6XmOWܤ*hPbU{Kˠo'0-a(Ut4w*M;cm!E3L5r@/Ƕ̣ HdGAb8G6X;,lOO#Hyú&Mnzߟh]$0e~QXKpZTPw~뀪:'0 $qR\x4ӞT'JU%ÊBUAEHN4*̙9Dus9:_π&7iŤj9\qݴzDkbD8cVBdL1XO/{zS'VwDIgs  n! t- N 8)[X1c2b=6MIKD [}IJ; -V.V}I&W>948滕T:/Kwxaµyֱmv{p77Jnd4 sm v/g-/|kf 4E7|6j7Vt?M{Q2ڗzqls9G˳L-<{y_ ž.=|&P>ZGK!zƌƁZ 1b"iZ+ Vj|k 2ۯς~Mzt8/ifz/<% %T$o]x(`90uL(xεq1 =ic=aҠ TY䄶+Ψh5cJil c3+bRg^=Rg n =X&b6.W⎭󹹹V&Z o+Rlrm 6Hr $zI31ZaH),q ݆?sh5Z`(3*hL nQ9P/`) NPB2Kpg.+y)AtIM\sX0d?$tHڦ(O eHw[>ؒls$dI -,RDq{p,㝿pN!ٍ؋ls=-^+k)`NMa^!`y TQ͍Q&,y [ 6ǫ5#T~Omo< 1}¬,* \rmP&qxޢ9qeFEBs#`I P v#X0M(Fq1F: lp6HhjRz\Tknlǝ5=a=m{s LOfkR Pu&UwF[]vg#PGcGmyΌZѳ _[O=~==ʣG|'5{ |vc] \۞i=bR-e^ʤ/0i/`埭4I'P.9i+J48Ka?\ݴ{7Su԰"FGDdH2+TF6S'$ŕ<"9zMw^I78(\ks.-V+T/oCzj $ovj)S8" s*PJȡHI&9[;:c]7S`yꞁ0E9; |Ԭvv-mWz5 M>hWp￀\&煖Bi P:([}p&AJKu[e@0yV 2;4ԊVag#fG-s<l̊s^\g1lV_d#[p.\?gU8$2FTCXZys"_ DbpP=tynMa^u!翿ή^+q)"o|zrQ2V?E'`Ldd:]("MnJך#/Aeh˗Nn#k[Xi#ls0_zo\=SRb{$"i&$ǢmvaKMϘ㯘PΏ4"r[Ni#Z$ݳ?LC¸x%SPXQ;;b ɟkvn1Lɋ_ghoЮd5%=nˋռxo6<_2L%EdNo3 z;2[Jʼn.' l4 9:RuaR9CK1ʼȦ1V]JZYۑ$e\лīaV^*XlV]⬟fL}3NK {mEoj%bOBR4Ag?iv1WM*H@jJ}J P-s->&2Ky ?wRouvxy9 xUv-5,DNIJMFOŞ-uI/`D.9i6L|2VN8U8}*N3)g)P2ý3)P&Gg&7ه2glIXa3$ye6ExXΊzN6UQ`>t8vrwP5Goћ`No&6n|| rzX+i\2@`M>Gr)D`oޜYW)"؞]qt^Tn"|+@m0EH9xv?GUt1"*y6,/j&Ivqϖ"@yBi*S}\^ww{BYt~f[b?;7@X!t5ƿB&2^&͠RP-R"?9DRYnC <3cwkj'-jVh*_b%8]x]^%u!j-P[RwV&3zVӨƊUu*kayBcȃ\?Vn{aY# P2|VGRnR)-E4릎Lanmu0pzñ' lN(aC; [ ,G.c/T[K޵qd_~10 LZ)R(Yդ(JVTx#@lUkkU۶Rä$2 xCaiq.Ƭ2ZW5hщaspqoAB8 0=ꑠ}.ޗ;8|Rq3ɽ7X.},> vy9<+fP57.WR̨Z 1d aTa{+A$ yd&X*nF)K<;Э1  s8bN0ts\J )Vࠤ欜D6ƥ d VQ܀'J [e SѝQ3F4GQF `UמU= 4|A\PS!HzTW&[$ti)hwWR}^R6᮱= 3/=kjEQ_U0Vr&)F∶igUV $ѩ-C^b-jIPUWuD$M&2ЙX폤Ahcthw,3V̪(N)JRvNC<{(0- lיxa;[Tir¬̷Z4. A\OLK/Pp4F=XzrmU28Ut)\]p1;`Ku :"/Q%* <@ "5-+T^0 T. U U:?8{ ċ @hEE@YR:[Zǣ&P܊`m@mKmueSU_}(yYQj2JRH`1>FO7oh9TXy%זXkxh(c#P. 6piA/hC*f7]/룫FYuEhe`EzV An!@*(S`qnj@pAs^ϠB.ds۝ V5u@XK4"iQihA4i298kkNj#:]Ad&JI>2@!~ЃA2AC("ep% ~ X(0",0Ljt#Ϣb@r",U`k6֜ß?r XVH,4NjTRp@Y x!m*gUn2+"ZzUP̤ ?&*#0n \n A`t y.nfj67NnQ)۶,` ڢGw@BÁf=0HSZPpR(uZtkHk sQS8%  dzz4|4 |殭@aRB^o*(9Q͡xy%Cy!:)QR(]&T*P=]Y<>~?U%9Q{Ga={'8 (''WrmJ@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 :/ _k^3Z^?w'\kt Ï9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@^'HBKr@mO{1N )  (!'V5r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9N 6ս$'4w/ pq!N"GN *b9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@_裭'Gڊ_O^h[M[esvCo ?In&*X r=$ 'q]˖3)e3f>˓p(W' Gf^TZ.Vyy7a'tp:*U wxߛwOjcT,[ gifӒ~@Oh_6͹yr W]KwFՇEjd~v~mzvlI1Ou_^51j@8Q2^/+e*n7mF5/I Ey3vjv<8TQĠ$˃Z 9J^' 4]q7ʺBUTb~š4lqO(>3D]mv}ntr(gp5WIyꙡ<}J>dv i;=tvp&!+SkR1~CFf\NSlI?o'>^x9;qcv~cI~]q=o!r_OjzxKl}zr1`oOoKeϗz;>2v˗hÇƋݲߍ~5~s-f ~U.g^7gHUˎ߭l ?}앪@Tv iYVƯP"r[('Nqb yt7cbyVmbJM"γ$۱n_[mf*j;^0[o۲ya1Onm9<+o?gtw77` oV%JtKͶ lJH(+zMǬ/.%]^fhU-n2o'jo-Y~9\F_^=>n!|eK\YZO\J%;7[fv>R.5_ )"Xbb`&<(9c=h8TG9vUt1[l5s۶^fF 2˶Tyu xUm*D^ q/aS(t+wP>{O\p19*&t yW3_t "a>>8Ew8Q2t]G&t$󅨏bְg^cyS}|HvEw5DSK2T6kˇʝV&r sbˎ'P#>NI։}XC0}1.ddT(v2\^;eաE! m݈aSz:n'#K8ω*xThK/^s^sa^sͧn-!mGN cɶcLQ*fV(hmPժ ѢZ舒li0) CNDZ],9Ð8?'׆[C֢?o(+h|\ןX?tOpʋPҠVx/ߚ#Wj7ȗԾbR$nWTh 0*&Xf1<윬\saIfY3."*ciO}"ڮt-؏tnX,TP~V,Eڶl!ml;g/Jj71Wj'GNUD;CG>/)P"1ϊhGJ9$W߻nhb/dB@ f g[U(_7^u+'1 jwˎǢښP졌olU.SYj&s;!;]s Ho\TRij늇S/4W%qX` S2 #0 /nDa:ap(ꗿY.  "vˏ"錈P;Sk cۡg!jCSob)JrWDqTVYpY=JU Pi뜨3QGzl{؁KyG*@ fkARv!&QA ua PDdrХ@7|3l1zVn\afO^+|օ5۰DW&|8Z` ` ǫτW%g0A*(aI @U%ȩ "|=TjU*@s\ԉUf 'Mhk/17WZЯfլaΪ5ǙO3X/*!3ALeM~* MV P9deL%𲇗=o~JgcӅ$)hN} sMGoI+UJVTx4H䴳FFC4R9XV[\.i%ZcKn'#v ^~t ]ֲxb髷l)Fn,IdZ18ӏ$wLK$fHxdR܁3J}|IE rl,76gC^lڪKɭf7Y QB*0)yH(xE1\s#SE,ΠLJ#?AʺmL"d\"f?֔PQ<-!hHT[ RV9\ir{5W%N'Q~{SܘK`A ș#81RIkH347W_Jd7سVKJk+ݷ6/? kz_HxFLF$!D,C-EB}4D{Li-ʥHP2##|W31<c0/D{y.)\2!pςh )D(D *yQ@$.}c"Fh8HIaҝQ1xΒj}TKŦVwLͯfHޟP<9u:f tg|SUf$Bx4>hҗr6EFp'T΄K9`F+4@ FGRCԌll|jʈ 2go.8~Lo_ =IlL":^o\MVWuSD)la>FahƓ/y*U%;!nМ6=..I婁R 1Y\ߛ>| 7@4v4L}d<Ե XǼeoV?VCku89jF庱7Z1cWϨ-F(4djxJM6pyෝa6=a6?&M,:E]э{ʚUR&f1Haۨ% RZY*J Og1;R cՠW@ߐgQa/Yneg]!;`奼Ymu3wW@nI5Wi"uqf+I SS4"rkB.֢(80J\p1ȀNFQtB@'J1"y4 \e}˜._<5O'x'5IQiP2i;J.@2pK[Ṵn/D z{FYM<M3nP+(v437yv[ٔghi 'Jg'+ cj̓THVyWS]Mgq#B( 6NxĸuAsSg)ͅ;"HIKQL-u% 'D w=)q)YL>iLq O,̍29.mn}>%p/|zb_8;dրr (daPN%99CSOl/ Iy%l%o(#P(Oѷ _qC HSQUʏĆNS\?ůK[fpg~Zp}Jkpi$Ul{_0¦k+ ӥ q`<cyld@._/q ;bgEx?]_A,vkGq k$M:G~aX0a/1gBOj5)7d WdzB˝MJ<7l8?9?_N?߿;?RO/?[OWP)pB=7 ДkvZ\uỌkJ.c5#ެ5qqwgC6gVmFŝ9B9Ȏ Ωki9[||Ž4@f!~0Xt":oVM?,D^lKy !vƾO"hmEx #A]xeaөNJopNДw L0ܫ8P2, qՃI1Vrz#-jb5GKT$W"Gj<$PBt$W )fJ"?I Y"H'r;{E!%sW}QGerkzWq7z\]XE%Nw_l/C.ٸВ+dtlP66()զ: ^qƍ]', $l1%U㾖[w<+^?)b&z\֠aH-̒ z%>H[_譎ڞ9g(Y ҏ'] Z~}Wh=v-6KbH k*[VE9ZXj[AmlݼW=s%25_ޢB-C,H&Ry<$FΜ k_="Q"Ĵ$]2Vz[5B-XiL AT|ρk#>Y>Eݫϐ4{/)ٚ*:ȯ&Xc]'Lb:V zm%'fOf$by[q?lܖ~'@\ke{g s-cY׽ѱW s.W;vpy>i+|2Y]M2\_NOgrkxEqҸK>n5D`z9ީG~8:Y^cI^z1M[w'a|up]|/X ep?]"ò8 u m\*ee=X7o]zoVNſpvr7|ʇUJ\{z֓1W}D6_:X=p3I}"Zʢ;u -RdZTYA 2?)p#ܕ]jҢpIp:Rɕ/{ghxyrg$5EZY4 ںiÓ{''F曇^zv3Qo{m ObHP7bPu#"V_=jס-z",N\Qj ׍ >/<_[-MKgóVD; &Hx)uqQv>{1|b'ϦUg_ބwVw;ٻ%.Ж^WLv߂G+O<ݚ>ޒtGvg-Zv-;˿[/ }y֜a'wwrˬ{'F2}'ˤaK2c;OK_$Ý:8nX<| fЏָe0ci0mν)0n qn {ϾD-3g;/8P%[r ye ^^?3~+vtS GOczDc!S%oGbz|i ˣ9pO37p~'Y7~k/iV ѯkT /9pS0ޚl<ػ6窻Yv$9|rźy >XƎ?(h^rw\ fÔkWHi|;_t܃pWijDf鋃LjDtzݝs_;ahlt .ec6AgdC\7Yǧ5_6ʾpӊ17;wgc)b8Rm ǣV5wՏߺT)=qu7>o濻}|:j<ϯ`5BSw^U'O3#@QC^HbnVGLD> <.j]mqnƟ<5J3/ }oWeM\TNّt{-mPo3A fOJ$;{ C1A"y>:g-F=6DZY'<5DqR:\1nǛH G#ֻ^SbW;#Gxnf)(''葠.s՛C6%W*8#6SAd*B}Hb9iYUm4Z:r#E)6xtr=ْU"?f l̉5Yݜ&k;fE<ȎP39dpH&OY -s;G;E>`uZiq43}t99/}lZ1K,1 LmX ufݡ8TGh:եQF:#e`4Z%?e0D@9XP[ a gY T9|ӌCQ2>&J6Cr( xLyvqlƧP˔H%e AM=- YZBU* *Au<Ø4pDQ7fak3 X*U8ݲ60(QF3ܳuX\b8܂8 nѰ!R-%CKdC (pm BǦ9XD,&(jl0M?4@@cS]F14z?/hWTVc#]TP4y@gJ3aګY,KL+f1*%jtB3>v}ê4D>o9D*'*кd@mXG  y$8ϙ8m,G/XH肸  * <@("5*2ȇI#ŲtiXѼ80{E21}ePV{uL< o 1*>X8+R~T}gӄXEޱQr8UV+Ai%A4D.ÜvGy'UM(Z鹎+ +xm2=\Oz1΁6gƚt+" pa Ac,9;k] mxVҵ6B ̓ 0@d,ʀGf8m,C  3-5Ռjʠ P6 hwțED]=By8M6`" nPB Ɩr2Z,U{ct֙t4DFjvRk-Go(p$Pf J،O6X=yooGx0M92ml|[<_MOat=?Z/c0I}@0u[<8z6jnQa-h ~v60,QbZsuYkd]<[#8feh3138c/ mgg톤D:(SiC?-nD*ss66Yc6\VNRID$bYjD)*.") ]0qfd LB *DN-4H H*p.{#ކ:7,@6ƾ$vMл7{۶BH29/n7@[4).i)UxRO[/KT,'n]̜ߌk&z/f^|];UN3"b|6$P.Sυ'O\ I/^H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! @qH Ȱ̟ Uų!$I ƍ# @$rƐBI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $B%HC?#2YlH]A sԩ@K# E@0mH! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! @EAyN$PQ0& ِ@E;yr@_" "Wr$BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $rHne? ɋ_jK]I@Pn jU;B C^ e *xU(|^83lw( 7}fJ"`hRԀIфj":s)T.p&f6("v}AO?ϓykJwoL'.|*vfPzI3 r$Be4s#ڃFY}I*yq'KY|ޕ~0@^zجS"/||z] };Pۯ^|G>>s@I"5VY(|<5uݿ\Z<;Xe͘})<321'{:ٺP]艑ϛ㮛MzŤ)4s@ʔ\Ƞ,~b*? qȰyvNȯ cgBrGޔ2̘{hQy7QrgAzka&vj x['D!,X91}hȜH%:jgƝE. ,AgE\MttTYهYO=Bi8;SvYAz^wK-ٯ$=L6 js R,A88.Ҽ`5EFYƲ98 XTh"zpd":93*I1L݁BUdz~K2}UUI.AL>ugϿzO$e|R~M˯^Kד9y2yy16n[(IӆTz}Mk{Xmw\$ɻ;} VOIzO~U$kR\sw2UF=?y8l"EO%geW =.KDQ'EV%-ҷꃿԩ,M@6ڨե$WZrָc?q1erM_7:Ff7MC{Gw.ZDfܦNQBR4n5ʎ(yBzq'!`ܢ`-, %eצwשFC;ɳI\1{mQ` 6(Q3wݬkǒBǜ(l~13?{@HJ[{ghp6c}=Izi||QRy -{ճQÓNsvRy;4HOGA1}8ȋUR]Gu׉}ڿW6f~O޾&h$M*x"YȜԃs%+"xݛsU8WCA)cf fZTe瞚]1VhG#M;R L@HfAG79ɸ/$ẽ?Z -$S&dlH9r;FZ݁=1FAyALy2IƗ0 $BZ*AUJ]< ! EV~8kUi~"ȔN>_lkZ,ww;WZX7Apcn۝g;۷j7 5F?7 pJi}n(/K⮭Φu:,"d&h8V˭1i \M3pf8pv6n=vԼ?ji`Ur]%㲞&>qK+eg }+t2PhUH V3Oq9hurIZA}_>^}:4 5ʹU M'Z Q<)KhWR[v*.oI 'x  S`]Ӑ˂r4cĵkX\8/Sk3/ CPx)/X>PLX^{dYNeԊIa C К0lakqr>%~=(cTv=. Vc˹x؊1&4 :pZf`JӒI@m jk>PV KA?n븂݁sciҢ#kJ,c_uzg=>< u5a `Z5fO䨾6ɳ-)Ÿ~5ګ[ZZ kAQyn9,w>s2 3/? TĈV-ckh[v؞Vi3e[X-<-(m4lu5=W慴Fzj[l% |$$af|\g'd23NZBTI<78n =XC fHCA|L@T9jPm-FKbnwkYVF[B^z  Ϩu%_((hQ;2U{`ik$)| /ր`x"-#i܀Zl[v{X;vT,bkc_[-"Zčgr™AS+CT9A  XOB80PH:j۸ޫEZv[+Mm.VdpZgkd_(Z"M=B KghȘSPG!^YU(W 1zǾPl~&,V՜]ϑ[W6BbL }l(0UgU %:O((~0LLDtT0 ursRlt3?ND6,ߔ邓;I}_w(&|xtFشy)q31k[xus cmD31kUFB}k:ڜtL[v6Drv ooۛ,-ټ>͵Y'o4hed7U/oO|r)/? Mn>w^]N$%fޘܣt{pkN?}uW' |rV"$Z>j oo.?J?Z23`dNo*_r>\r=?-M*R RK_T0YdG.|w/ 'n#Jׇ%PWjr(JxHΫ & LLx5LhҏG O|䧺զLw>MFe5uf~of[j)n KNwo6`ؿ݌!_ ϻq~H# 4e,x!4f*5(ht;^ʔdʇ.v֜\D+RaN(D*3ܻqvRATbOSRJPgPh^Pj]s0>Jz[04ˌQU ut9VsV(%EUq]k{gpkE\`X~NZVK?kKO嶌I-ޔ XhBe 9’,y$)N,-¤0TiR)RqW3~WФ}(>`?{?`eD0XNScA,eJ1YeeЙ20AMױdpTJkMͩ`I L2qN ,x4k$*(!'Mײt:dPֲch´3!e@ ݛI&.l:91qfZ2.UBx4|"g+CxbUxفkډA Lne.r?>!cU΂4WyA0TKR`|n`^΅ Y^hTfm)Zv:񞷏袰ٳg;0eS`jtYWPJ ۼw'6 zG$E-lJ= Ϡ^ngho~h;٧R~/'m96I؟7]G~4)|FSLI Y(hsF(\x%y{]vc7 ZtaF r-f $f!1Js(I.#䶝h"lVbg2ag :7a֏oFw4̟Ϻ0` ,i2k1z#c00VU\~Jpξ=[` ,E,*ŜEB4R(,s,wK-m5.<0\>0+0xz>/w&fSqnk}XXxR%GR%EXH$!;e0Bȵ1J}N Va+n^1U.id& J&쳍GLTBf/ &⑐T1,Tt,B,qhبSp%'=HL1 ]3v֝ Z>]C~砝 !pT GwsPF-3&\dtd l$!'!ƻ$%%;.sI /P`O7g|v~{Ӫzz4EdR/Db,+2yEeNۆ5O'ue]9Y?3k/B *I3i 3pxi-7RF w&Kz/-º-ZdvSp(`N um snɎRJ*ƌ3$H^yryc>3F"cIBdad8jjAyLzObo M10_ o\1 2R)f(K}$ag_C.ij8\. ֟_g<]0G},"4nXΫֺ7{U]چc_\ŘֿM6(mj5I7O3bWcGOV_z9X⬴x4%6o sDpK6 4CIj)Χ##H2 eo_ק\tsFN\MZbT&a\<)8oG{Bb~,1(R楹Gᄾkbzy('7KMV\(}#Ծ%.?".")qg%iESi3bb:_;]tx=-&߷0O_>6ܿZb{DQӌqɿso -wafZAĿ^[~thVp]eƘshgleQ Ʋ>LduCu*=wu]x.zC=3yptp³ cH=V6Y}1=d-s%d ʱB/"da6ge\TFB:PGd"h[o@l a7?<,-{3Đcd8i(TL9s Rjdhn^!ՍگĨ͍5W6ؕzӆ^i%oU/銡TD_''$4 X'osw#uFj!#F?9CYwb޾ݦ%zi'dr 幍_9I)ZAW*prmuɄk댩}2 %Ik#>&' y}r'Te0B ^]@,ud bG&A:9̝B5B c#lh,,|[3QĀ!c1T;S .ر_G+?Ah%aNpgjY|];ӺأAG[CGf8{<2 5Z^Uz [h}{mv yoY~v.Bn8u#gIi aw'n~i @ SpNԧC7JFR<j(Qc.<L.x>PG~<;mR.YB26Ym0nP)z.37^fT%å)iƮ/V/>vx2ER[S.A vJL)MwŪ4E]Ë z^0C:>aѷ-<,)]wNMIʈ;[ȼ6$SeNݞ'f%='wΊo>bsjiOZ_`<-}dœ kFK,iHfv^+DB*K,:2]" eG.v͙AHCWK5  ol@3^TXM 'Kr,bؼlh`/@rs+ғAii/B[`XV}uw-eFwAMa6G/rk>:-r.$k}2^# $ǒCҚ;Y P+s.,+ u)< T${\%96\71D,`$ɑ $%dnx':w_yl&T :O4ݨEzbQcuo 3phQtBvq1"*ML?gOB4=Ҵ .JK ]: nf.B*<(y,$I" zAV;\RnT}ք#6 )S0dM_ qlZ=Jsvr:KK7y+4P)/̤m绛9u˝I5injOsx&% 9PS)?"Hn2I'~urNjd? m=7iKin`=X*VU^I7wКEwWA8Ż7i@eOcS\[kuU2%woo.h.ZPV.caBLra̓?.h8%$N7jfif;/Ώg>:R.G}ppt<Ya#(6⼉pF;d%Z[BysK'mfͬN_R!5|<~ddp~)V9ChuR:\BJÒgT}1!핫?}ӽkZTjoTќiev4/?~y9n?/ӪG#pBDm`]:`сkMׇ7oztE޼iapUO=Mڵ]>rK7\XYSŗ$ίs fl:-QT(Vq+> kL0+z+԰D$6*x GV] mp]?Bv}۶dq2".|ƬS*X+K|[F+^&lt`c:%Oͺ5tqjMAmh*bz"}!Hkxf$'6T'P;Eٜx&RsQ߃W=t3jǣ{QXWDa݀牸CyՉ ǤfQK+Q,iO#( _{P&%2Q`Bvͫ A["Ath2ؘR2mvF%L%K:s `'S&Wgs\le ug2&ųc.;;hёeցg]rLEd/R^[7aoeZ۝.ڈC8ID{D? `dqW2Ew`FG=Xh>v]=8gg ˂KP)$0 q&1%9YP^BِbXgaL)OΐFc,$ Hgݹ),Z" -ҦTÕ*jRE&)Ta཰lCX4? B>ˌw)`3#AG^2xÿN_ޫgtu6AvJb 6CDs{w=cabk$k/ƫٟ.hD׆)Scr)Y{>f@r.Т%VE!3AHptd~xk2->SW%vyI'/_`sr^. <5|:RPTl֠2&o42k:&+|&`"Xd,'P o';Co$͞[(@W<7 bb2k܀o{) .EVhoo&xFT⇯Uԋ UWaJϣ CRޛ+u}?+WWzW}_K_R懛OAXݹfsT|>;m)?f~hKemˏʂݝ/+}f4KaCBr&V5WXTX2q S! 7w8'{Wd\>MDCNe(آmr>SS؟ &\t$!b2O3BOJy#l*"1IL!XPWμ{,<6HAfM.b0 t!8mBaϛ_0N!]9a6*3!#d`##CL@-lzaup|O&ħiv]1U A7??Se;gdBHՁ @xl0CB'&o/ޑ*!ݹ${k(ų_k1 Q$ƨ2 sQ!$Łw CVsD޲Jc/խ'ݶeԻO_cs\W ,'6͍9q/_&_&ӫIτb։çHP'HP?uG^jW㶹ouEΖjOӭat#iP:PT^5TRgf]d.]WOSlUx;:A?Zvmw ]j?SϽmg+,H鄪aCDȐP"e}r`>6E@mTU6Kd :r!HKHe'R_Y'4oV3TwQ蘢qPƻ̳=u:h}gޟ۰g ?O>_ Ap>/D}l׋1~ż}Kl|qr:uɕj~ёDg%u#q:Y~Fjjw,>r1߯<,U0FJBal6i~Wcwex6j7K!ve%AzPޮ+`i)Pî3ʭ}r/~;ı-Ynя#+GރJ Eh\S]夻?.,atuEqM兪)UaɛNԛ_sƧسiU`}.{bQ{kܥ-%C\FFi!SN ICEY9ﭫT60 A0 |l4't!IepBUHx gcUf`V7逴M/]Hd3̈́VS+kbCP(9iEdL&4f0HpѺt=1OxzRO3*_#>ڭ.&=3ֺ,7<ĝN)Bٸ(Qa1cjNoQ WT` L=3 [C`>EOZt&Ђ,?)*%вm׀r5%?Ri~:3oW{%j|`>m~37d'eJ0@4Vgt3~:Y8q\(o/qVp#Fǣ|{O?ut2Zw?k8`W_۾ OꪣY_mONg\bHdNz; S1W檗z=^d)2^<_pt:/vZ^ q1H-@?``vLʻay#Q} 0gB+z)KZ)qHI\yЮUuThD*C eIb$-0#ḋV(AP )K$)P5$Oe [#gtoO Ay(Y$7e3B~snϋw 2'rɨ'y/t:O'c??2)~4Gч C PdU ,<X*B7kt>h_{Dn8lYf8,yoBL! (tU\]p9ϛXu .u`|/#fPL2Nm*7ћSQrdl짔񈨪A䘥~*^gE|<^.^j4ǷaIÜ9q.8KR{Lh=:VPJT~8}nz 5@;]R CVL]9YQ#zS*G.rnb~ y/cڮvmylFټ畖!Slotwxβ`.m7ӯJi,)8ne|6|+.c64oԾEp2 &EyKǬy謁/b TDH[0reEHZD.j[: S 7 d{{G |#̧0g)ʨb$T $rAM*ZgHڋ"ur)\ІHK3&H|2 Y]ٿݢo*6ZϜf9PcYy5^}n~| ? 2ق7î믅?cmkkCi L\%gQb%%/?;'-.eW^RjKH%C^JgcB9o 8°-$X&(匷NjҜ18HS^S|"H:]p/ϗ^OČ`17rRA+b\E#WRI,&LSmաQW:AXP+QfVE8,v]_Hwm,т<.XQZftZ8LKz6-N'pV2A|Ā/x=M+O2KA S.HT6Jjeq!]p6`Q1+M$YlͬM<A6ƜN4pPp6V;EI1 ~.M-cS^@r̩+LIϗxVӱo?ϕ{[ܦ]s\7E LY;L 1ꜢEFS<ђ9w aκIp^Jߝ u e@}\FɃLUN0BhEB!J.( X1TN( <#,P(]0h|0N+3y$)5RlXݱ˰v໣鹎}3θ !\`WA&;tW*EbX߅rf[y纳ޠ~\?⻷~]7O_|ݻ ۿx7xW:J@'$qGpo追q)kkho>4װвY|qMSnw帕B󋓵 ϋ~3&>t΋T?iyfɫcPW\;Җ󍏟uqtE(D*T[Fq h-Q  OQ^ü} E%h*cQGOwaO[ըuRzcsh`^)D ulHBKЩume%Eȋ$Jdl-OBNQ $ŌAIOhIw 5Q!l'*BC,BC/ kqug k <^ -am% 3/e;^{ÖL./-St|@uB*^b7iQUtDB/(<[iT`SCۨ-eUNiK=lC Jk&DJ&.F! Z{/Qr[yt),Zs<"4 F QDOdR9TPͤL2(X 2㹊9)ek߫.:渭}hg)QF DRBYTU r/T&< ӤW ^V G;ɣ^@,u*Lg{!Üֆ>] rf'F>$GJM.!Z^:lS[4"tyS>/[ZWUR K}䈣c@vA盺|E1v~ 8޼Dm'[lw HnLLl|%ɨIJ#u4i;J.@2p(I ϥ5wۊ1#lkTZ{C &ϗw.õbJCl~Ӣ~C~ %'LQKo TeN3SQ@uLJEN&+_/r<2%h/iJì,A:YzgS O/J295/`t1kO@pݻہщWwYo ht_Zӛ+_>3|?o~BWM~y'ߢOBቈ?@FM /YV3KH-%OopBACѥc><xYIii^>Vn^m2#ڡcp3_x뜡DB0*QYíN*hkZ A8 Ri,A>@lSfO駢ψnNz_9#x=x7]Biy8ᡓ d|:_<_ƾ9éOtalCP<,YU4L4ʏvWÛ5:Ey4=7b,ƫXj;7 \''mgz^v|nbq;3NqbHV1{UП^tuζE͊ޜ̖#3W{< ܩFuN 꼺qOEYM׆wgp9ֵL= F~ʍ:ƶrnWd7=ߣlor;Wy_Ȓ^nY{v"K:{7[I Qt+rdZG6O'+Jl]bjuY=.`Ei f?M;`MߞcY/'Ӥ=m;d۪jO繌{d؏s媄@Q9O? y>-DeJĔ21TԝA1.V-EYWldC D MQ)=nZ&/%n>FGυV()5khe:%* *7F+BƧuW,C CSf+M],| eTx>?btRE>ýAϴo9a#R7M-><(}.׾tx{cZ]Yu˵%J\ &YelV68;BPKB賥L}~içO%ǁ-P̖$E)).];#eI: xiZ1!x S%:)Iv}$ģOѸw3#|l?ZPHHR vEoYeV t!rag`I*'F6:Xנ1*&J1Ȥ%#KrgkJV- 8î>//ٙ[x`x(vl-u؋gPC} lox]C:x/P9E6UbJ2% #t^3L({K6Y$|9 ڽ)U7iM2a4}]?j=wԾcgI#MaAx}ΆހJCaM/sY}s;XTV9 n^ iQk3X,+$/C9eK@<.n0̱7xq5osaT\:[nwZ,'Rh{PM7Fz|*^R$GfY#j&H* 3rLRHA JTUG>}G~Hcy&'b'\qʱYC~jOEǹAFr"bKR2$ =wU96^`+:g|yu{өzz4EdR/Db,+2yEeNˆ5OS geݟ8? kW/B *I3i 3783}(-Y7!@Z:o~ȮRO }+ТCf5CC Ae7pNu +JMЅWI&8&!將6}.f/T)jњ$DFLCC%Z==JTC.&BɣK(}pnqӺx\淔P*Yo ei$:š;'WðLo1NU{iZW_PWvf[W3:t~?<".{]nJLmRNB,NBWQ̑_F8d|}boO?OeM"D#9zʠ$.ݯQ"MF4 oFDzXdvORܓ)I1H)uiC$> 3:*qoEL*ns츍`L |YRM*BFg}D2r蔕D gS^~Ws<랕˛ TMK\L6ş^=Q3 u3eZˉAh$U>Ʉil|2 %I{#> y}v'TeD ^j,ud bG&A:9Ua!! 64A>i୍(bVz1q獯 ժ g{e((q_6)`^IZVS;'ֻ/zHd^qmP m^6FtVF)(#mD{d ͮ/oj5P %gTrcٽKvvvw_W65uQi ւ|.Z̍RFbh>{//+.|AF%bJ!k>U)fyv6ۤ22]joxel`>ՠR\fn.hJHU_+qoV¯Jen%O)r'\2 h \s )2IhD˂R\hή'{*}㒎7-BDTSS2b6(2o A&ɔ1;hSwgYIfɽ;O\]c,xH+,^۵yluxre`{hR%ɬk%Ѓ^HwEGKY!nU9S>ijp)#gQC q !4idBEͫ gK{'[ CC[+PGfdsF]W]k\9ײ֢w60GFf95de9>.ac!HiͽܽIN_, uL*=ZP$X|qeFhHHFuu{< dMQ)}$2֞",#xA<LFr s6SsQi g2e鹶m)pU \RQGWj36s|R@$IQx R]Az Cwĥܚn,`_2u CֲY:[G+'j0\.SNt<rrS?k5LJ}$txrJbzsx9Y<`x́b. YED8pI:g_&WҙmfŒCZ[9X.0e:0mf1&dg -qGbv6v1Łkp95}\Vffwg g_5ZPV.ѱMLry`k+Ni I+4S}4X{݃H[wߧW~j/:}a1r9( ON3+rfWEF?__I^Nh1gi=2Փa=nxg7nVTvGVG\e'ξ'zsx>pjJpgmu]v=| :)4ұe(M_OH{A X--VU&wp4<ËSZoO?|Ͽ?~<n?ǟӮ-U %c'Oynu_~޾kൺ] tj>l·|~5ha$ߏI(/sb6SN=sz8pA krYQ%U!`@5vn&;ٹ7`7?B=]2zN8LBY xA[[7c)K%-BLhAG/ALJcq:1f'fsGi}9aJe6J{dYc  1knZ>5<3> U% WDģ*IvXME{:`Ԟga-bYX7da݁ɸRqu)Y撼F:R%bj x&%2Q`Jw%Ē&%J)#(@*DL)ddL@9ϓ.e. x=876{j]@jep5V|7/cwįtXi;C&FKp4&~' <*'Qc*f%X1.SF U(3*Hy $6.^٪/l#%{i+h!D`9d h:=3$F9T|ՆS$Hٓ:AZBOI4ҫR)G aBQ-HTF" y?Mv ~8:nr:h״ /RM|2hF ٻ6d llЃൽoa;AS"L\iYUERgT^>|\U"Ѻ50`2kMYi'P˔A_?帅4j)@\ >/UɳjlwDXwSķ:=%+VY07{ {ZpIU{^7$5dgkԮ uz\3:}}!ok0pg) ΘDs}}?#WT-%|.‰7.?$\#rME9;kzyõy6ђe9X*b58ŢrFmbhPyclJ SV\$o'F[SweFEzGp\cew[EPKhw&T/cpg_Kp5hF*3!BU%A'&^b";c;ɱ6$Qs-9mL}׋פJdVCI{ֺ؅C!fҀE68R$#Y@SJg)+޶[w0Ǩv#F_tN/|\n6x߇zB+5 rioUS_o/Mwņbjҍ}(ԫ:L܉&ө7w MjU@2z'iR@H<R]$lEnK'u0ENxX\uD z)f{2(6Ҟ9IKt8r0-{xbbP1Ij<[֑a4qGkZ )$ښբ:(q6.x^dOAˋ+/0W' Gtw/ NG`BuI#7Z}M'wv."w_.oH '˩(R%XW!q\R'7 yvyqоkJ\m3^ɣ 1NPp`XEbDԚ+i%QJf X+0!D!e!xh k5f,`ZFL&Z u֜ .ךCEtwP 6:R"0fV8Ұ(y9dU#J *j`(ia O> ̂HB9U H3ﭨ5aE ,l]Sov69ىv.*ٟvhܷ_ua{ tGfeg*zDWYZcҭH8Hh}@^a#>RT[L2 %t #LOp).}Ԗx93Tݖr8AS=9ҩP%x(6"-h,bcrEnCԚzɕWL{AF nl'J0H+TyVFHHqb"Fb3[XڱG: H źo֜ LXc:ڤdW(k]\'Xu#Ta-gEĽ6y19,X kdncCIǮPlnLX* E//-\# |{vxU2 x ~3o[4yab):ώ>t}N3t+RV!TPqeU|cnf>AI'髱[}~鿈g?aPe< :#9{Ό#q'N$Ƚ ׽n~ӑ7bɴPE% aJɢ%zdK߇M,C `Nn\BU{s/}*#J UK\C'OJBwh-*{ d?*Ѡb+EF|ށm65%h g[{(o6?$k.IcwY1@bҹVä9+gҴU; tfLrJ͙"<%I cr%1{&t:hs枉#{n(L'4ڮ3RڪJBl/Mq!0u"s\fv̍s[2gsn5|nCĬld-)3$I' )wN)9|UM󛠅b1`̭%8j]WD1u-Svn[u+tt>~6 lE-6((d*,o.fZMoa@ڽHrjYOG痞.9n)Mx`d&+m2U.9B ,O6y *c`cWWhp1R>Y=B祲w< B:'q(!-C ȧ@R|hK`O"v\+Is U1Tn&;n:Yiè'$Zt"Z5\SbK-"Ƭk&U,x9yv 3SbqΨ93+9C2ʚgSG ٩-?D (H[ B9>,+@&Д1тFQ4`pH֜ =;]^l Զ'8EK9Ͻt/&V#(AJ1TڇT]1rPR@oe\L ( xP.Wm2]{Y+Z;9]RdՃ-ɣt$ʜgj{H_x`KXK3EjIZojR%Yde6X{ kzߊ ՒQ j|K[% RooUۣbp)dtVf `[ a}FKoϦݡeB]@@ 9F`Li(] %#*dlR RR'AP$FZkb 8j 4"#SUYmW_PkR"JLehS˜WWI`4kML#L z_) tBd86Y% |/hR{Q(4@l~+n k$ghA33W:T Rot|7ԣMu0r4c2: 4W6'|,Bx'Ѻ30 ky[EGdBZ1fL[ 6 Ќh.B?R*u*VRV+e`cWLh\U* ZbhU%@I-ltb֝1M7q4]ZS wc{ƹڐK]:چN]|.lP|āU&)1\bk`RU[YĚǐxz%}Dlivz=A3x#. voxҌtJ-pdZ]h6:&4f4twS ̮r;[MCL֧U.aɾmʒ )qepAsAlαeL=!Z )EMeYZ=lvr6_@c˽Ĺ5K=r/G>-R#VyGi9UnmRe\dm%N3%V/h>3k j /^Q[n7]w&% lvF!DNXllQ.ȁ+FYkTJ@JYT5F@,&QiA#e|κ:*1|;κ@;۽ qPT9+Ckm-5bJ+P*{SdI]PZۛ@Ox~~fШ` #;λCH" bщP-jb{ jɞ'yXk k%T9sEò !*%-b0,@;.cvӉ4fp?AkGfLNuB2H'_ETFsO#""[+ז[E/vȶo!Q^Z⍯*:@VSL JP"d|v<.ؘg'w:iȶ֡;[g689LGo$KƋKBK O dgJ6AK=Y2f;O&GrJcլGkT:;f"nT5w|(}rJe#>{SSz@I!ʦBAVhBں²Kp 66BOu peͷP_&em/t8w. ߖA^9ݛ3Wz7Χ0Vn j`>aX_U\wBn9tKȽ. Z{vLײvJ ^yHuY)bv2uo:(OGMdZѰ4 ԣwY4H~]sJmW"2VoPiԬBY>͘t`ַ1h_Yd]|PdѲ"V}#9Эy6!rQչC|"uY SB_GejMĹ?Dv -6>ҝ>.D3OԖ5oݲPBM715}MLת> @78?c,7N<=O&g1}};əX(RG+qh)vnmYEn—Fέ-e ]%l 7[VQGLҕd&KZWM 92{W5)J$&S `-RV0@)Kvy`ͼ;].A?;oH}PqMJ5#iEML9VEH-vIdtސe +ɾI,%8X@ "G%#Q3dIqHAr"$*ʝ 8Hrx=_x4.+[NC7=*;:,61N&+ةr‘ZҌ$6jKnpWR!lVgk%P C#{hIܦBgZҡ`n(eϱkaK=vM+^vl]g^{`B"H'ZNcp\-zZ1S3/0?bz1'r`{, Lk֔GQALj{C7qKRxnG={ĽGs;]Z}Gl(.A 4LU2&Ͼ*JWhB"յTR&˾2#0imD() M}xa\%Eq~.g9轨>fU*WBLjmʺl#*հOwvӎm!u?|$`CE)/Ղe.{4eُϔ x:M7 |o`%\Y~?]oYYɚ1^Caג_J_ ++ b$WPD,?7gKi:/ʞ z٢]G!.Xeu:"qY#,Zӿ_?~G}$Х _n~UZU(6f@^?;a/N<,O段dgO?Η3{PBkrUVU \QV[=z{4 x+WA [bƊW 1b`ث| y..V߸Rr"6]E],s<[傶D)g+1H.Ty+*k)cXrgV! Tc@N:2|Eb54 uu_׀vKʭj⽀ޝo}6y *Ȇ1&*dIP5RS{! ](5#l鷔~em7pwr~L˭_$72]?ly|<'NNa.cBSv1TFSȚa_[Vd{%[uKbOBmR-_TU0`12d@WiC}Lh)[CjږlJp`0V :Zثa"P@>:GL@9R0gdj 3݅?aNj-CFs[lxuŜȲTBH^UICҒNkHcX9ȱIGߚK!oyyW<BFrjbH6ZMYM-VKv%Cfu <7,z;Xn{fLҢOVsn.Zٯ}&118ʓHXϥ7Py*&dU^*+p>m;_<_'c]yZԮujiB6 E q$2Чݻ/o$ _k*3I8W=áHICdS"G4{zzzJr:%8T0ICB+ʈa"-`AB 9~ܒ||t|WlR&@;0(.waz l4yGiVđJsXr5EZXwz^]`Wy9(gKR$j5D$ͻ4v8;_(ҹ\~GfѠ`յ|xa*mMO݊ ӝ<.P N3LdqigC]zQev\{lq5 mPGLf5 4.$+&]7 os_$d 13uKQN e!&Gj0r= }A)m+ ^_x-s]}ȑA.d.M[O ,M @:iIr֎,xU*hx5L-H\9V6P,TZjg] P_hc"1Mo3 RHF%S69k5IP:^N,+2-w )r HeD lSfpX~ ??'䂐ꍟhWVi]BmyW#\t2紑|>eg؜@U|M(mrfC|xYK_5ҼZ&d~ޮQ<}c 7ǷbZu<_̆ej>G02A|jy՛Etӛ} u` Q 7ʐbo ɬƨE3XLGez}NɽW9pUN]>~aqO3΋!^X-a嵬{],&Ng>mToWi3B]ߡ-!mBCPvHԮX^Q)ݚpo֦W65ۃA?hT|欓.!SYvˮI[vM%ҹWZn\ԗBcd nz-/J@e*9mK<.wMWLy<^p\mye>=mɓ5zayv8#|T %BS-,#T'v4։ <06ِ{ {(XԿغu_JܡT'P{Mz 'x_]a2,KL؊J֯a3A\Z FUq˅I朻YRm ޽(x.Ho}s} @/Rg-ךROmZ 1AσOZ#U!2^#v/dzC]+ ZG65yڹ_NQFPH*Xa&Zh[<mmy'z]AE1,|X.bRV :R;#xz!kq D[I2;;r\"{HNGSjjN)WCpZQgBB`찇+vor`gn?_/Vps]Z-`yq2,2nwu:ZNBljn}yݻ>k;xg?|wvۼ}swcz ľs3OyU8ysrړ`7'{Q`ԊQe ay[P|ZQj6T+wϳuOe\?Zvf>; lzR$g7I]]vۨ#Og%~;EO3 " \D>: Akp&$p^s.J.xb';]9\c)4*j =Ȁt@ ,u6(X438SD[b,H-.UZ)}VH36ڙ\)8 So o-lCOfdiM9eb/_׸ԽW3tņ|i{]~^>Q^rJ{@'(\J##"(MFd%vbEWC \c9)ї"]ZP82.$4EAtRE̒dbICh.)A5d/EftT.}@t1*Tx5Ύzo/@ͺ= BvQidݛA>bTJr!6|x5F dR4J Tۗ" C~f&81< 6 PB Oe62jmlV3 H0@bP_?i{mJɚRJXb%!\**@"g>*)PjQpTQb2WQ[YdQeAx'O0pE.ʉ͕2h`j hdP$Nt2]pGv3u==?+&"*-|]P@)|32+$EgBҞk'eJX9^8G9X%)hG1 W;Xa`Mm1ؤn,ɉO79 HB77nnrŀ1\Mty[8v;vYYJrxUIh$~x(O*g{R Я6P{ڭ4 Bx5+?5oOQ`V-9~f6w &› m\wy~U) j6Euu"t0qY Abɱۼy.~A8i>Þ@²FAƨ#VDel1A B%2Đ(ۘq B )EV㴊' .*ef#H=ECGlgqM&N_lx-yEN}Qy_wkHX/!D_@ͧr9"$c eHu`uŠ.//$ђD4]tt%F$Y#("Ujj{;UV(>dnp"e}Œ;+P,Ej1eg뮇hZ})`Sb]u4ÿ:ƻcMMwQ$T"C8:ds[ Hb6(J1;KҘ;#YJ%u2R/Т_Z8ς-+*OfQSb[|Q+^Tա;a_gzeq^﯉gPO NUۮF>8v꒪ @ "_W@ut364ⷳUT(o$zqA{79aG49MW4dk"o}szS2%狿V=;8;>D.բz X].w_`F |S#BJSUsW78ه~?l0ژ^>&e<~@ymIZ_凛ƾ(:JF.rU9 o`$ˍ+P[gmhb#2iJM % |5<[!VYz}BoEDpwٙb+%ŨAL,}uA飪F $,d 52VlLIvg1pހ5(pKIv hatqHGl%s.`\) o~4(˳ 9C*tj0F һO$k{jj B[1& LM.z,cwpJ.`Cd0PAqWf1,6OƻJLo|Qrd/AXV Y>ړNup&h uE XoHSAp\ħr}^}nA;뉜.iP[h &Y9aW{/6jo} }[oS*<SB*!tE%֬~:F0 њdFʂa]}Kk˦ݢe!E] ? h R u` JJbQcӠ bs53GZdJr| amdtEXIsEb܁d0q>TAv۩,)"ޙ"@) [nवfdE6:WqU5"wrJܫ'Pyb\׈baA,h" g"SϿtt혎Yf:=X"Z7cf}n!pH}zPpO !Rѱ bMݏ]:ZR֥ZTYE |^ukM9k"2paQ1FIk ED@I5d BcĹӰM 6kP|h[;c01xѳYb{|oȿJ 8v~à S)bV֕Z4&o =Q{-f?jۓp/Y>.Z0$ , Ц1Ӣ0 ؄D;l!$ DPF|6W2wgN[#3` HH 62BX "L@ͯRZmF([ D`CP_5IQMsCwԴ9OBӦfQ%/WN0,>qyB\~KrWiӍNyY{Wf_{Uf3R|2ݜGCwY;]{NpӍ7_>s}[>I_[Y2<=4 yf`_xe[˸ga%KθgUO5sk]+鈏*aC:Q|inzoB|43by/i#kʂxN"ƿj^.y]3ʭ% #uoT>=oad®͚[ RFz6r%_/>6o]dᑭ~uUA̹e7V^PԯNBK<+ɪ8$kCy~wpoNf]T{8e}?R0I%žʯ,^)+hn^~SӍ^wM +2󦴐ZS:+I݅v/-{Ϛ'hz:ǭQXf}PF>b"Igȵ/C,%ZmU[605AE,JUS N{;lS֓Yjs1He։` ʬj%eL%CCxC!W6]LJH#($Ȼ+c.!/YCf^)j:V<͈/O St] 11`sRcT >D,C띎mmy/x'"A[mkv vO`然h (8wILܼRiw6@E잼j wCD!b9$t3H%}eNY"`\>tu(\\F6kkso#5i5ywD振;6 Msɽ^7ٻyf2}7P >Z=zü+E{9tRS~7, \[p6{Q>L2nJw GTKYl]KWU'h)m+Sx%1Dي?e&)菵C@b5 YӆJ>ğ&@}r>HfӨI8ze E1&bA[4 M"AKEƈDBU+MY~atヷ̓[k8h;0#&?JxO;4H#}Wl 6MFhҵwrS(h-bh^um d];yy,ߖ3%%   MqI'a&qЮ*78ɢEUH) x̒Sу=P%C(1xqмCsǥM6AE[o& @m+PP1N gIHFEK1tAi&*gA(E$ E16e ((Nz)td%ArA4,$s3,eGڡɽh.=y}ߐ1yvߪ{d6r~;ռi,_?Zsꮩ"26vZS$$ Þ\U$iQ2⠖:,9.\. $J.9(s`J)[* -`n*F}m-|T[>yEX/f'ut66w삶:7ugrk>+IœM:sBFe 'bvV$L :*Qrv]`cd l*ˮ0 FKg, K_RQnh5b(qnq:+.XF}j!R*M@#TNLɥX+*la,-lVq|Ծ*Or0g}ރ7 =[8 sWUYro!F0mzaN5b$|MAQDLG$t\AEd\+52!; ڮ!cwN0;~o`wYW[_JolG{&@ ĩmu#/i5m2Lt`僆 e*0Ɂ;x/-⥷juE)"^%m2Y7*1{aw:8 :6{.H)QkC*C>&$IrμTC LFMsdc^Rx#j}6_["Nl:μkbURj7m6`=1.A5V8C'|ل1~ ^69`vc+8NsrL&Zt' $^JI# Z"1Uow9;q{.sW̛gㄜG)˄nb,I4J:y wT :1_uF)YU#MeuUbf)=<&yJppF iN־uyi{.Kx(zc:vm'p-vNv-U4Y%hRHJL@tBk3i^q)DRw>nIʷ AWuhyTʹawfv=6JwП5'h.P`,,Lߢ *ЅANě_V*O_pK-Ktgp do$4J(s;,bDW eesj4W_?KMxhxZMLNe&,Ñ+!|` 0jSRĥ2:A"D"|ی1A0<`pcP+f $>E$qM.3Kx!cQi 9h n K" 2< VJgq)1Z x EF.9GX|oHI$}#9=-W%!pqU']xm3**M=Ljz1u˲|'ӦEbr"b=&JMo(@)r^B9a<"S I 6ͤg;odfYj-xˑD- S:I8{f깞4ydPLPǀ@ZJJFd k̺ -:xXp(`QE}kvc@jXD⅁Fwr>7/ƙ21iK&#3ZG E0FOh{cܢuQH-A_?(EJ$ 0”.@6YzKR;`GgYZm!{۝ \ p6Z}[gf.?隿iG]u՟K_k2{Pډڿ0POaolGmtDCi|j{>x?gqYHc>~)#﬏BGuT娳O}5]Ǿ+1-ζk'0qv2]^w]񒚟ՇD :܀CFb=%|8]^-ʈM-?h<dwRC(CemD.j]=0yy RGأ5nz6mZvjZaaRm:M7P>lG?^iKS zxa˅yd,q'ޫ`_aoA;q4EDa]1Q7ET' F"H-W=F1+1k{l(Mb}+Ļ5Ǒc]z?CǑcet].-_#Z~j$vpjXx܋T]jw"}-۝c&`[խeǭ˾5i_}>ݑ`zMYk;*]=0u t)1#2w0Z;&WCO'Q^m{kL=IƆLRnq|qMv_X]9u`YY~qbZfZ[FýS@mN7؛ڦ6<1Im=Uj[r{tFmqXNeMlS@F'm@+[l;`ŽNv Km!`z]>]* =XqnȘ+up (!hd%TtlY[SB9 so'|s bPCaf&Sr iN)|,<)i$\H~^ف!W-mo<﹜xbq዗U8-E$`HuL.RbMiE}RК_e`be!n&S~r1$|Є\YO @g#V1e^ %5Ȅw2L0AZXǒEnmIc=OGJMNgTjN }sLDU-@/Tۘ϶c~>Ԧ\r 奋}H:u4ڶj5<\uBbJ~brT !$p`flAdL84dǢrV&6h4/7AT )0Ee:C*/M6dὡa$FaJڒo:QAV{\JUɋ/g>Τmc1g) lf/3j%0+P^W&X0^xI^ | -K2GiRGOfRu !j/l~UYoK)zǀ;,Fg29rD6vG!HU w 56A.hd)HiAQvEHό}`lsxfoG]&n)o1:].Krq\Ueigp*ڕ|z5?_`Lh1h*r9]xMÚGQ{]g3~LGŀ֚/AA{j$0 !|Pf3cwP5H@:ͮ'\e`q'Q 7.)FdWTDkEhʍg.KdK{j:?8ְitsד-r[뺫ϴE6&V|6Zz\Zo4JwBn=?.MBCsז(t#z;`-";)Us,jagϖm:ƶ?dPZWMCY\;Av}{u>¹V=oܿJ-VO]$o-l;?q[hpxYoA۽5W_5=\{_Xmem9X۞V<6˙Ts`'\9`b5(*`a{A'}Nw=!r%uh pb41д58x6 =Z{6%>Z\by>"yEl1@VG<71 lmꍜ Բw`VVX뉫dmɲ*BS#E%;ϝmY;؀&U'CGǥm2.JJ+@ #\=̑"02z 2SUw=ʝ[MzLuꚛ/bnL]f7_4WJm$MOh"M뫊rb~!:hmZZ wo>^}n6<}vooNܽN0M&noY7kxռhk\;r$Hy٧f|ʀFkP}^-z\}0AJ[-+cYJ x%k B^`ȁfetf u+9;ʫs4P߈~EW]xrON,)EL<8<" `PKM0Y`Y1 Y*ՠՒcբ7Q^YwxXzˠ#uW$wkjLԺdv^'@F*$1FSv=~Wʋ#KMnOy ,-p.*o zbI\'g= jnZQDs]΀o(m?{lU{P7Tb*D|:O+<@\u?> V!\A`4VY(o rozL5xc$|MAQ-QFtp@X8"@7**+kmfv?w"r?_["N+ڶ؆XE(DdSGkfOBpdF'|9A%`4{Ɩ*D*{8!$mr*E}>@%!7FDbsw⬗,:\njre٫gZ{X_) .Q]m,جcdE?mH:A{N I9$%l.vmÞStW^l{IYS,늃J2ZeljlV#;;O$9Ybp3_b~Y%n6p?| ߯P:Aq=Co_S-/c{ qK^_gl?i~\cCNAS5eFlk2FElt){E6AUG\ >:oFoܝ=v6B{0_GbZ0զIgcakUQ]ҍaO~o:c~~{zp_lo_?dzYZ鵵{b͙er x!\3R=߁>yO~ZW|0PV| 8I ݦ]^z_HX/N얟V4]Aa\|41?[~jXjV-+7NW{jNGbY'ZlPW6-޸"2#pls_ߝw.Ôx\T>?A! D̤L0E|ŋ?䦒BʓXBֶ~b:(j>Us={ߡ[Fq:n^jn; r|r`N2_H}f"ǖ@(ZBo&ɚs$J^[օ]4,K%z?;  wڳg7I4ߖ\Cq|R?VRbӓ.JEݹFjgɫ`2grF.N5hrʤCubөXRUmSՔjV{U9N ]͝mb)Fkѵh鶎\8ޝ+5˾;)詚Lf0) ;MqO k^1'!gş>|jq5AKTU˝TR9 S $]1ڪTgף(]w0aW#1LSdmtG7s!x 6Z2 Z׸6׆s 5J"4k;i^,HU.C2+nʧhl`-|E+DnJ} /Rꨱ 0ijp3 ޞ6duGގ@)OGPhnj 8rJcg`jb x=+p-iQcGp~R7`HԁW J,d,].@kk#K38PI@ ͑ձ3ƺz2?s%P\- U-L5V942a=g%{C,x@ %8_;wMAQlz2v4[5]6jb4)*LV\ 4NS8SXmd2zbig] T ++#q3L @ X "aY0(3]c'͝:g3| ݚAn iD\ZU53&nPdrnie/f*ʌYdD1&@Tؼ*BNZC"/f3 ϰBs/YA\ e]éA>Z$ >:%ƀ:)M<ʓ`ӦtdYlAJCREUF">,1g]EQj۪-=ѓf AnK)Av%)%ܱ-tt T(W|`cƒݤ%$˽eN}yA0 "253\\"Τ 2SD2Q*(=(D; C;SZ5zP&LJBȺ 3A7 ?ˤB DRq`k>Z\?X-3, IC55eVo jzr j oR@ Qk K#,A)y1jNRv`yat9?,Ά״v,$99Z aUw&@3 --z(mӿ4m'UCŨ]Cb]k Q^yHY=yp4vMf cM`ʳw =1>.4CNQUÁvCPB^"úbCjra4D1_%dǬ<`ւMJԂH20ĦqIJS't\- 5HY&*?u튅@Q YNPn 0P5Kd!>Zm_2 )+:\,^gCPsWf9 gRh&+f8ޭ uh/7 kO?F i{ dRӯvRv(e3U%OI|> *x+G_"fM$ce~o 8f2_-&It(@E*ٻU[ǻsO+8H/Vj=\QAg(JDvP%JZt]āJh ܛsDxD_Y `r9bvJ#ڢM 9hYAn^-TǦ9u0Ŧ a2sok~L@㇛ɐ~</o8oZ9b3}Er[|įqg}ap:ysKoIr2L6g|+/zW"93 .By7!X %Sgc;.Z٫ʐ?%EJ,2#"ty&A멼}$ 8~e^FXe;tG+U0"MT$[}l]_ZC7&A?שH? Bvu+R4;λ \z)@\xyMuw*[x-нvm+wJj*!Y]>a|\/2Nttb^<XV~=[;*GKYcuJ[}N*2ԑ~ɵ-3P.[mC޼٧yڼߧyM~=/e[{H3_Kn41}Qvzfڤi2oU˧m~xY]־]o<~fگ_H1\ p1?4T v+ %g˯77$.ZRơlˆdEݢƆiqm__sv9_-vT\{D ۇYɟn:f(;r ?,D F匂+sIkfRc+1 }o؄]_Ţ0]zw|Gׯy']9SŠlR99Y=>A˫veᚺ_.[|pp>4!B]F+aĭ}gI5/FL1qǹ<βV~=seWR&u&Nc[=إs๶RIcV?d[?4K g_m |%*Q鵪 ЮZn^9r/޲C{T5]4'XypzO7M|='o}lShV@9GU7|ůo:(ea 07s"eOPNPNP9< t*6˘4^ *JG+"C[ Wg4Q.S$澎/f\/%5sub~=3W ^1bн~$fo!$+~co7VƊX+~co7VƊX+~co7VƊX+~co7VƊX+~co7VƊX+~co7V~ %9!E<:A k{:H) "_"r]d#c4ʱUh=x &aܵ*d"G:up ` ZNULdZ΃%ʠD$r LPX<8*hFFHEz5q ݮП|lc}-[c#Z_!00UO6WrZ/6G9l`s#6G9l`s#6G9l`s#6G9l`s#6G9l`s#6G9l` 6L6נA\& v 6`s蹬6m"Zi|waC+ɸFt./rk1D>)i|A .p収~} tTQ2f |8$F  fE"Y'睑*غFIEu5q~cAL֋=d3຅1X֠?i!^:X߇0j'xl<0^jx`[ӮCF߷9"NpF)'8iI#Nq҈F4'8iI#Nq҈F4'8iI#Nq҈F4'8iI#Nq҈F4'8iI#Nq҈F4'8iI#Nq҈~8ic8 x{c 90I#e ! 4ŷYšے_`UeBh>t]Y[ƤQs6.w:^5HooJ]t~+nxϵ `M(w3`Yz߃ :Ɂ|FFs༼͗ijapy5ާv}05dsx4߆AVA ^\_ȱ; @fCo!2o@2]olo¬G1#ßggúkuckJd\hW7[1>ֽzV4rƞ# {?Htoi@K>9#Rdy4Cuppo~]>n/\?Zب"aP;tPUFVBJb+ψSeo=4wz'ٖ˦|tֿu~SRG;Yۤ 26J|AYAy9ɁC]U|=Ĺ[f>w#p7f7qY0<ܙ=k9q+UC%,s%htra ,ޠ:a-D9K٥J\\QsϭiԪ hǒM F=^jtڥ0h6{Uw7Vi #*sH:tդJb\nd܈k."`wt۾nmem,]xm;CmO&#_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_!_۩\{S5;nk.a× /sPI}bv'0%^8Ē CpA[:)ͭל᯸^Oݖa nW\ v{GY6 TނTi)q7:*`vFNsBL@CjK"mU$jCHj܏D-kNN}_z$GrA{==;/}fp9 iC'题$n_.z=n-F`uI AkLt.dv:3+JdAF]LQcU$(jn5͝q*zvdu`3rQ.~K<$3ّas$:V"Ilg52/*Y5qԳN Ԭ&*\~ ]F$%F( *%#TC9d^"I%,U \}.PoN?f9)(uH.Il\ȼ)I0-\w6@Sg'4WOOov蜝M,mZH)Y)>/ I[Sg-.UQ_^)L_hkYFLBgm|+rf-2B\*&LAJk$Sܩ{!c7`➤~P.AIY*^Y2SNg93ڇhT \e ,aL;cxT N[cmM(R`q4pI :)Ͱ̸QZ;ZXT4G*R]E%Y>[F?5X }1Cb ]_$ rA:IfpQ ֒ư 3N5$'x!@H*^U^@I-sq^xە'ꟼ-(E|J|tb~buG҇uKt`d_~ޯ[^:_-j)R0l:}݃IF5>{aSB”M[7%vẟ&W.|}9>~f8[#r1}z'v1K"h.]7?ڳ-ԓ!{r}O盺aQDnfy jSЭQ0eN>} hg7Ԥ8^Cnuӳr&^GVixs0 ֗#XrS"G6I [X įM]t<~-wo緔o} 87 y?=Lnu_o~ؾkju ͺFlӵߦ_[7F}Ρ)ozP }v?igNhR9rXjMbO FMޢ7.p%Ā).U(>Ɲ>Ҫ-OZr׮7^nlŘj97~ۖpv83G_A!h,< 8L<.ˍ~?\d1)KKYHc6Z^8 L J֋t*Bu+0>u{E?Q!طxA,[6H[mT)1^p X$k9vE$D%J)#eSSr lwp'&\=KR73_X&Ζ̍5;]o+Aμ:[+hBvރDeW%GQLmLX$BLkWZyIV&sLVQr͉r}?c/kŽpXt=ӥ,5\FSoxB%Kˬ˺e1W F8m%_r^ѽvE[v'KZ}q'ϙOR㼒11)F=)fѱk*XJ{hAA($H*RM)8i!L9.PBUQĬTcSSYEREff6.P B S*aa$J #Nɠ| eI@i;g`_VE eY#SIpI,3Ϳ*_ko 6KٷgJP (0gsk_O۫_|4Y o QsZX GiY:~4]Ն\U[qKhp<9\[ۍϷe0GKK+($l7簷*F툽K'pw~Z\7*P߸uA(FmTYeMhA']1(M M2`!98rT*;Iw[>Mo r_nrOP:Ff98wط<}ăEhc$jic+(uѴ* Fc*'J@`a'M@jvE$ pMaRFLDŲPBf*-U`90i ^-JH%yi˵)1%c#dH" jT`$2IH$%.5s҅ B:-U`I!\QGa[䡌Dj%|U3d~#FM_6лHeSh\ߛ?Q~7v?hF%߼n(R?96qЌRhh:jx{|9m~P$98hg)43pOVnکzsIsw?nӶ9 X Mzqŏ˜@^f78,+I,Z:PP =n~9MKZs]mT+\ONFo# »f.]vwYۿwxҺBd6Lߞ ˺uyt߸eoP\yS5eٔDJ{<EuFr˚;f:,Zlx$XV ZެٿLZbCRy,,~39d kvViؼYh#[փnsKP*Fڦ|p0?L|5b3f{dҲd;~/$ٖU,QwdWE=<JrX\[&㠾5e08VL[m< +;ps/" G> ?|??~ǠCwVGR +1'*vGKVQ$(Η.֝^D|quJ{Ťd_\TqQ&dVZ)QVHDcHF#jKca1ua< ,@>oZeFn |'T{/qWƍqx ~,hv~4ys~ G瑓σw&ڎ|ގsWBb; m6d F2.0ƅA=Ř'Os!p:_ݣIJupPakd}{8 𷹷鼫 QlVBY\wO׷oEsЮ2Wތy9G".) U>ڌ6e[]M\u+ϲj0,M𧒑ҏNdMXK=sEOH!Up_T9!Dc/6pgoC#å-fT8&. oy7R)fUPFNFhaht{giB)#A0eQ49X=t螓;cVcߵ1Nm}꿬)}o:XMRKFܒi2yal~H#L+P-cJS#@jDk.2n}egw]ۚ?{<"8o; $Rq]ؐ\%f3;^ʾ.'vn!)^6]^R -*卭@qUi\@ƐH B1/[9M(CMRP V{IM|vHyTh$D*C ʁO)F 3'yP/֏odMy0vkpgIlb F22A~5Zh^৆}Ft]{ v,vkpos1fJfXm(tэM's>~3BJ͸gO6e {!=xdۮ%k_Z]]Ed&#Vͩ8>=y:Er}wc:kU)Q]lln:ESغS݌?]wons\oity潇mͦA;7iz|w3JlݹYɆW7.2S(7N{tdQo(Ze}xOmkﶽlﴽNӒA ]| QZS^oCz fl>FGυV()5khu:%* hIx }tJ(<{Mz1>fWzOb7;={!~{_Ի7+W~)ӧVK@$E%)=]{eI: ,Jj<$tr@kCJuR*ce}q :)UiMg2aVr"ͽM h|=Ȼş<#^_֑-@*_RaI#t;-aiHT RV9\i$7e-'8Fq3{ʟZ\(_4H|E)c \!g6orH'͢tJ.{l\h:7(W}P 3fogb 1; l4" %rj .ꃧ%u$C`JkW.EʅpDȲ C  Fqh/<+K&$@Z|!HCAH9H.E*H\r~TfLP3ə9,Y֗NV;;l N s_ML_P<9u7{a=Qlcg[?|+gܽlB2lm|Q#TI/E'Dr l'|)ƅl5R ~2//~Mlƣ2ӕmNu*9g Y bI"B66XQ^b?^T)_>p[k7ٸ:}qZcN=^p窯K뚡T' Ο9 X?'s7~4}>M NxG[d-ZV[3N9e<Ñ|y _Qe@gD eTмvzV嘩/gc2z0pY# qk`\ݜpb1)|:h1 gƣ*=C޺Ix~7/T[yNw:RwLrA?5<_֫fƚ bLxUY; Dxtq!/f"b HZZ\hѨr@IvOdr21ztR ʐE6 K^fi(B KPP9$:/< PT$ UN:g6 ŞZoٜqIE2eܗ_}btbk9!T '%d7{=:B#{4tGhC"Hn[$L} #@*b7rifO("md^巾,=>@hmHEt׈ ȧsJ?v.B{p\lLH]=8F{zZ5F~z\6dqFTWu(cHQŐhP6TwuCUY/L֜"8O:"MrXXBhj>00l{"IX}lƧLhkL%Qˠ99(lHJl0ʒq@$QhDh/9{ TUuP{[ylW#wەBp!ܑ>|rEKRRF6b^o>ņYDOflng,6|pW,-}SX߷5ފ-rlDKmzPTZgpٖ$% S`O6`m22J4ڔXq"E_O3$&ِ# V)?{Ρ;6qkky}@͓;Dw敯L.Һh4q5c^tEqY&#ۃ`SF-3Tl6cT$")0[|vd3F0hnoswpkAϮ>dƄhY`f%4c`K9( ۫QRI4഍X!7{Z/IItZ,+|;xճ֝ ӫ^-xJQ7$=e5C1t~A$25Ei<2($}/E3?X(*Um@-L` p·Y,BIVoP+v`'-wi}vU)h\{.~s#d|_R;SvFE9L#9>378 KS-M(R!@ T!f@Ѫ£5IeàH*6ߧD8φ#tS3Kץϥ~s_8g$x1:lVg{m>#:o&+K:#KWeICk"w2MN}*ZEЊۺ~\3'lʋ{9PSB%?bHJa:xyz>v~oAAl#$U+ʛ6 u͟y@FpxX]b0IȔЌc;>7J#f*חKE7`QI!nJ k;?.xK^BNn}\yyfV]ٝ?\_x nFi4.ǣUrdWE,h/;Fx94~Xגk[x{K'뚑k16**X 3'#8Ked|y16G';mU[]i^ŵ ^i ("n q =e]έϺ~kcd`yO9^%#hiENclc^>n.cJ%SʐIU9*P-`G]݃B`a#z -<;CX)Q ALhkēкX飪gĭ/;Yɰ 6NE,&/$byͨ>C8[Pq<W RG\9%ekG<Ϟ!ghQ6NZl8[_\ A햌dVFkM\&l(\*( m"(@#4ݜxקn_zlh-ɐP˅ڕB;H4{°"𙈜ԞtT"B`EA{Q2zrE5ܺuF3] H\swvFOb,9ţOj{˼7ao^+ pȺ@QkV?qB4XY!H9;awc9Ѵ;!8KQXYXAHH 6Q QYR& \\W\ "l\u>x0&2:08V\ºI̚۳֝/U!|79K;, 0֣4lyb"ΕQ`0^ErFثt3(n{p1H.( Em:,Py&y҉j^&x[%Zgt& {!16C{&ǜUB[_W?#GSP+:a\)6*Z OJ '&V'O`'?m-U:63$R)+X\s ɺT$nK&}&\d&\NΗ.g,|i,2j# GZ(X]dBYO^Ic坣]OYkW<]oz2xֲvvoɾqzK"}|TZu%XɛB w;aG @Pfw78SH8#6B9kR=i, CA 6!1*$3%UVF|&XC J&rN $(6(@YBπHHNsU9\V V2TT,x(# ~f6ꤩL1L?4-}e`tJuQh}Pɋմ%N0..V?|.ݒ\4yLNǴ |}Rޫku}7/kdo{F(5ͧQZݹk:ܴsf}YٓpUw\=oo^mDV 7N"'[MMH\fu-,OZJ3ۇ%^,X}IDݒ׍*w+wxxwVެO^DBDy* + OwƪDKWWXTz->?{WƑʔܦ#;U庺sT|${s5EiIʶvkgH8lɔüX2g8Fn46kS;Jx̯- <&d9ji]&'wbTn Kp?}j2xn TJ P'Qb.xw}z;WZL2z" {EuLriHȝЅ D0 EQwLJ\AK\ kP0iD~UjCV&`AĘLTʃ CWAFg q7uq\1h{Z:HK|$Vvd} ʻNq1MEfe;kTWtԱd"FYb":ia)ULkFqBWMuA9#:v@:vYŻWSOD\m- ߮9n} ; TK` a`ߏ?Y#%Dg +s7S!t|9LE@R\Xd$=6+1VIL>^|O>\ں[foMz׺w%V!"ur)D Z 1Ax'0sBn%]֭'abuu6S04b[⋪⨚}>~F9pZULnpj绦ϣLڅ}JP=JPw %^߹SIMXRӊ%笝 OKfVӽR}% jQrp96-9~ÑPB oTʹu& t C ]͞xOvG;BMrPSHi}F5>" -@D#EwWIR:H`LT% Q9?ټBB%.v[ҽ'(r5}'eGL$x*ϭ7G+> +j?jZk hR?ONrHlx誀։j.{=z?׎X^;qr(y;R1DA&(^q'hT'S"TFjpܱђ9<u !Fi&ƀpuP!"e` QYíN[P1rzn] ه*8]vP Y\* NE`E0hdҡIVp d,Qcd 5 #iK 2O}\ktD"v[QYQqţ[lWpy}(eNen{%ڧt|ݬw) L(9~+|چ'弊\;,߽TBܢ9W+JLr,Q˕.y,J" :' =s+w)Q(2* )\>مF@)C^{-!$XH]JB *n)&XXW)$cW,ApAQ*{U/ڋɿyt뉟'oosd#N$H: " ը"x@L* Z`p(\ew# !{60D8˰ˆ-1 alGIMR>GVh79]Ry(]L:vEmY콕5$DIxȠe ))ErAY"7+py4(taBճiI%*0<b :L^ȴW8|X cPxXx:TǮ #:"ﭙRP@ꠅ `dVYPA:8)HDc" HaDt#jK戋ŤcW<ԅPaIa&דiN/Ž>`fss=?>F #Mڰ0·2Htҡhzru3.4.$]~]Oo*jU./ yVM>./osdhhh߷tӛ&ET?]jv0t~j:׎&- o^Wm˫q[AVXX՟ꮑ RsA_8N 0ԜsW kTmlHu*ۗ2!`p}fV.SYάѲol!ZWbH$>ɩd,;,EN#Igdeq;'O9I=} ' dN突͸G|ػ:Pr+~D3yx[GɄ̬ ~v)zN8FJqWqȽnΔvOvJNODjyJ}&zϥdQq_Ƕ详f]9% Y_OFȅ/M_O? ~xV{<7B0R򃿛mb^Ty[~F17sM-2e.>7i:vp~Zuy7Of~k)ǰǐ7(R l ?[IlDe9h{U:klk ݛE;˼=ji?p $!W$Oח?2IN%׊1!i0Ó?Ex]s1E%e=%N^ږ4vIaWδd ~1wW`e*Z`m-ַMzJ1-ֹIIL1xK&Acv> yG4'\釺1IS_X'V @FRj":}nGg:@s xk&=X~k%6yѽnŁp}I)_>R[dFksBp VNv*wZcb(usw_p+R~Jv^z AnMnlLڑ]\8̭o O}ɨXw9 iǸ /.M_{L~jx\em='&T85VxiM#<;ֹ֯05]YYqߵ(oJx e &J%(AV,FY; Q D.k0gVe:߆f>+m vcUz ck0 M {;Twc}Y~1KzjRBu՗,=ժφ7~oqBPlxiw.˶mVgOx>nf qZoɘ !Vt7L}nS9oy v7NNawFCF l ([gS("f}#ݙ>tُ^nN0FIpj+shvL޶qXkQz5(_-X({Rv- ak()n =W5{}>L{wJBWbaH Чrtbq)dκl50*^Gx '{*#3+m5`G&(s>F=hˈY Ogj\Fv5,-Q]md?^лfr?GTpmSu)_f}odx <;q}0S 4$R˅m H;JigַYy06׽hJPZ/Ngׯ6ݲ;7vK8/U L5!7/?vXO*QJHwqy3- ""? ZwK뎷o֜ab$_y񺉹]r[LnFm /O]'L+>5VY|ew(^L3ށ-zxؒ7+۴&sNj!{ geB'-LvFaєI;8 QT,YW& Ebȣ*ڈʤ7)@!`oջX-Ҹ>kP s1V1E$zP2 d< 0BDJCH1'-t:K& USB EsTePP`60L"Y!9T0xHS)è|ՁcfgdX."zY|Rs<+f@57.WR̨Z! cѾ@vXA J <<(ZGD⦪jI#sHx"fG̨aB0 ?%RTQ6% L+, q)9nLWn_iddKΨDC@(  VWjԑU"y{x)(`~E AOU!B]Rl-sKK sJRVzxIل,FT0ȋE]( 㫢 J DpP()46Ja<| $:xe( WE7 JΘ @<=n[ncn#e[Ǭ*$c HQIbYj9rGy}2L ya]]4uNaڄqwJo]:lSGB$a1%T4mK[W28Ut)\]Hh)7VSc2(3(v5-ȗXP*ЃcF>@ $5-+T^eV***6VYѸy{ ȋ ":t@2GM8ڂ2Ύd P?`<9Q<|qYQ–*`9%^V0N+<}pk~> 斸sR ("Xkp`(c#P. g^s@ QMހw2ttrQV]k)dfw Z%tMކFL3uz5|< hwX@zT̨ ƌZNmJ'0%֔YK;l,댚$frP%$`?Aj؁0"ep% ~` Ü@#܁?U2ӉTIXs M;C:b8K4YF%Gj3Ko^5 RYU۫ ^`VAc-B@̤簡P$y͸t*2 cŐ%V`&k1ҘwwvNiW>p]S1< Kd0um1*ܪ #I`d*mtnLvWMKQbԺ`XC]kpy(Y5X`4vfc\`,`Wó=7fܵxS9njj>Gls mA*BwPzTaH>Jk$=|P@z;`}VFu&id zuEr\O.A"Nu63 ;b{C!DE HbQ<T5uclep?{ ڰ(RVGϊSZ]FʄȍL #RcZ |,$3)%&}*` WMIB1X>GYYr&r9p}N>l.2MgmIG >w27ow 5zFrGgfvo궁;ONZ#@n퉋iot?.r~^\u9..%/Ͼw~>ޭw~ȲMFiP\ż9w8׭_đG9 >QysG~/vE ګCwx{A񊖸N(?+pе [[˓lv.[, ]- 5~x [#N6ςݿ_]Nzn~/.߰ρ_K.rz'=|\5bH˫Spͣ?z̍x>1.r>/K\m4 nv5c0]m>[b;.|9}_~xx@knR&?xտCw^]huqH>>Ñb19:@L۝¼tVlQ*u1 1>8#;U C  j5"' 1~F z<>nۀs8|v)i@ֲ--pTSҖrr &SrmKpw=zV\s̶P[ ;:ȓ O\Mwnz}};jjwrp> ~wotm=F(J_ْFPrQwd>= ֫-] lo Jt^SUMmK'U|pC=q-.*Wr(Ǿ)87b_=ggso vg~}j<}Sep W⹃!\W.yPƔ Ls>j0LXFpM<@KRRHo, \1*di͡H+ꇯB%Zbmo4SF$,c@7V49lZh}۲[ ^;zK#b$^Lgylm'Fojl\2StTqօYN^tO7.͓L:F'[5i_:SLp%GgM݇trg+ӣ2_gX JXi,ʛmb֕m_37ĀLoVo͏|_wE%o-+'gO%Q~鰵{rr:=Vxm[;trnm]o7WJkyM},Ch xq@Xs4ћJKEh.f ?YM'cC#Y68G:F 9)n2"1L8:fbYɧDw1GنӜ*Q94гJAjԮlRBP^7O8NjWnʎ?t~ +Mj_睍&M;W/_퇗/?18~w 0@$c&`ACOwzXcho9.COj>7 Q\I&o.?/g%H?)N5sqR t"t;>ꚸ袧T)("^XB}> z .fpm6U,qZZ1*["7?B}Udb)28v/۬::!?%'Kjd htAAϝf9&fc,5t _#dk^l@YňY*<1$`S*LF><"31TcggdcFĕvǍ9Iw0j}Wg.X8`݀WOqKW>NXİүћM+9`R-%*\ m*(fj8`N1 224d$H\$ɡzmpkdl̹T xBZD=?DsS?3ށ(K"]Q=J}LDp̗虵*)KX*(*&䄮})mGm3[ `)O|& 'le u]Odt͏뇹t w~4te (iEJdoydj,YoEUO[[E[v'Dc=J RHTV#テlKR5.Y~^Jiwص/$ .R R1 9 s)Sjm=(e\Q@w=MaVaK=F1#ٔP ƭ:dH@%b u&rs߫}tn9f,sBi20Ym *pByn#L+_WZ2>Zj' @m>O߆m̑=` +/O?=GAI䡵LVm· +t69-YrX23<( Gc"#8iɖrU 5NMf]&x>H\3#2J9S2m^7>}zx<^twpEd`$1 C2:Xޅ*»@o$[[희,gȹ^ u Jw $GUJ5$$,jhRX "PS\ɜL& ff Ik>1$jA88>rw(rɥ DLX"\* Pa&Y8w{QEqTs5L2K CN/hJn:Y`:$b9JAӏaLW!JNrZM/" ,|K;GV"8a(DNZ<"=(;K"pEH#fBS<*a=|!|PR*vv~65R4&Q$UF! K.Jj(2@08Y `1 F&sjܾEv] Rؖێ8Prysw}xWv%͋< ,_bL>W_?nd#6}NT@?HY-KҮ6s+Ȩm{[[&x-SDg"3v%rkKZCBx>94-"D-⨸`sf\" 2li% Bde GԐe6XQ"C>Ǖ⒘3˜;]\S+j1FKPkG6ۉu7vr4V0Q ƈ`F%2ZT2-3)Our@Ō5I@vDG^DQ#$Hk  " V#No{OsM|tw\V}.Z͖(Ok0|@7^{*>nJ>U_J A{%)ךLOȒ06@2FEǦ)OMYrAL.䉺74sgRR5c5rn׌J5]X3Յ.{]pEQyEA}©׺]ɟ7hx<}7u\cR FYY&%ܒK&̲f* %"s 8\uO#H(b$H!EQR)L:Вpdu/tY4mN#jsIdb֮jmavckR =3 0w,䔳֬3d̑'M0>$z> \8"+:J4k0E-Pب 8|HdTs>e ʢwE#Vm5^#5p63ƅٕ`U.#-GHw2㥊0QgiTU#:Fm$wYN) )DPUHPFz XJ5rUҋ]]%ESY/^M"PGƚm A%ʂ)pqBv{zqWa5Vn@Ydy_# fw6\R)ǹEx'y?ZRîc/Bã+ww)l1\j羵966t'J.v;;1'AL]qܒ/"{b9imMQ2mJ3iO:c38r8rJ T‘e,2#Mr>io4.(cdn"ZˀG I@9w h3R_CSF#dI0^ƜT9ZӲPAzlbmiSQI ΰ, Mw'ɿv6|uMOgJyÙo_k^WM7F-ҳgwxGdwA.)6)ׇ7z5)kә_J#!lx,5h\gӝ\z)jJ<`@WqOgv'(2K%+*AĔPќ[ ,:%eX<CnC" \2֔ E6H w%W4ƳNKuZYI 'Nٔ4}_!bՔ-]|xW*Vb^jFc1 1v8+#(2axij,&H*\\"Aͽ1Z!㶹ʸiz F!51͊U RYU4c,TqVEDMXclro*F^=*wXހ ic,ȗID)LX^Xz;" Pto3d!o8mJwk WӒ+-_wP}8Oض7z<' 'vpad1b['X΄0`ODF%_i|lyK-ًC-1( (yva JccX*ksE MTې ˆ^FcD2Yr6U2""& 5#`Ҁh@>;>1錝Q|2?d)U"vtp[3eg:s;\7BKpզ^ڋ̑\YJ%88Ml 17`c0/ˆT\J*oSXⰚWh5Z`(3*h2:ڀmTKp'u;%x-:&9Bdzf><{[;x)*v,_Dzt8T"fnR͌ ϥ&g\ osDyE_@Pj|sꮬlˡڷ>.W`ovJ-*$^K-bЖo"J̤ݒn .ֱZS`SDD#X`^a4UTsc%bԀ 9v{ 6b$o|9 :)mP&@ۛ0d0,YlQp -%ava(dQ J!bHn9Vq7WzF(cgtgƘ t@ff%FSuxtJc]1+$Tv/'J6}+#rW04\X̑ ;pG#s* 'R)!S0RxkhW1z+S Z︶)MGLL}>(XNX !"=AЌRRA8);Ŝ(a.{;;jYoR4]]n @jǗzjՙxWQvi^`CF-Mv W<!!L.!ʚy NcKE¢6^\ڃYxQ3~SCvHmvڋLifwOSS\u<p,ahA39_mqkdTO|禢(> a` }Wf6Dv7pT Sfb 6i6IH R]ɽ2@u{Kb&5Ŵ2QUMdh5}’ēN. fbsgsz䚝3߼_LK(yQNHqdN 2^()|HIyL7ő/Ksh0 zjjhŜE"H1!'cF %vJ72/fr,F;[VW}Fq7 jPJԏϓWLS?堘Mʹq&g_BypB iȐșv>W̃2ك4{0JmCȆ{Q>Dj+J/EyEh+B=V*]NTz̓O ZDl ha[0 ^1팝]Kd ]OПCǛmz|`_Hٝ6A4i /ZLgxHMi0Hm׬BN`aͭ1gIjry㝧=TjUJ#2 hؑJ͝`Q#OiTliSNHYQn(Pʍ"VFudXDcMcEb"%>팝P4tR Ymp))Tq&)m3_g²eCWJ` #U|$4"2Ĉqrȝ5Ș^:X垼t/\<ΐ02h";;煥 ]*AQ#X2(:o;cƌL"^ˈiDk45[!jIg hsCY%\GBjlb!aDvBύ8C=lM#}pxr1DwdNBE6eg7fJӫ0g9PfpVҦOo&S/03rI<ӞdpADV.0ZE|dq>fԕ%/&Ek(}d[O(}(!GJ!TN24!`+^D1Go ]RFl .J&d? z Pu{J>KPfUಳެZ]:cSD|@3fUi&fxfN1>Qyn\z-.UBCwlv$ C5LaɍU|bYq dm:iRja(n(b.4(wP^hy=†C_ӶyPxķZ#϶:f8tV?m{".0I뜴ID$5C|O7Qe,:u~?.\c`bҹ12@ńלܷㅵOEg^/YI.&P>ZGK!zƌƁZ 1b"iZ+IӋε>- Qdsu^^Ho7C_jy8> O >YmRwW7OݨņO^q{t+^M}Oz ]:-lnYPÒ^Rip+bEvu( Q|J%?@`E~gy;0ΌvAAA(;3(eQ=c P\`% ܼzy,㛩 9 )l:W =3wfrG=z%S+" ])г~6ȊO2yVm,$P {v e%>(֬7ϛfh5 0!J4$3`4fjG]iPu%Hd:L0t ,W3|Dq5XBۉ>}{d%Z@rFϵ,wJÃ&nCy`D0#*D`8Z0F81dLmVd:+)<|m`0 iSrJC7EMVL`_dQ.>Z?2?Ίa G]w߼[}.{]ݻ9 o |RWZ,`SD+*Nr3B'8Œ'iIeCUl ܚ 893F4魛aV~fo7JQч曛 '0PJCJVS4)Cl=E }],KV*+}W|.H $ Ţ^I_Ŗ4-2e˞gd7]]ͮ*Ǭ ٸ!Z!adl?/dz9cُwkl3x7cɼڢ3=i%4L#ATVﹽK}UEɬ?~V-`4Yc{:8-&7uZS;->#(L,#DaV$ XtAUWTtud# C;/C{/ٓ;**Jt MAb3FK3VB,X&sSE jAZրR)eRQ 5 )m4"`I'06VfcT?ꐕukj-R fv"hK+d$*lZ[f]N)x2 @Y|]H|R0Z[,.~!vp9?-ezIz*`.Ug>DG?GJ׹X8F)bz~0 =$Q`g?Qa쐢܀Up|ם1I=˾ 14u"K~y7n(hIct2U(hOzYGtP$&}c$ar#kKVݾTwEխLRY.&K؆ ytǣ篣y:/f:f$l͞ xj~{_cق0@ouus]gF 󺥞K^XZsPh5tRq11/I6_fo'W]m0N+I{+Kgom9rܐ{ekHoecDY&ˬO2t)@5CHktIc)gwPԮ.c5o_~] YwJ1崭b+A=>JPJgAEZ`n֢)ZScHo:A"@eB%&a@"|CflY[f(vW_ݴڽ'W?lL}|cSW-Ƭmپ&fT/nС5NBɚv YcudDdܞ^0Ki'IYa&/!P0yskh5R(Z'5cq ~g¦Ԕ.5_Y 67zޕҟߞs菔R[`z=3)0%ya. G{:C:7رJ?'a j9`el5n!ϓPqoXV< ͸0.ev"YRd E.c dpJJ$3-ь?݌RJ{ڱ'5Dz"8b!]QGȜE,+K~z`<T lq=|~;6ܕۆWkaEiM3$]?!۰z;!w%H""ϗmb6"]"?56q~HbȢn^6m7,0~<7˫>9[^~%dKB5o! 0{Ci7ėk={F=QLGR-M#ϝЫ7Me`q>/W(]r7兗7#"=4#ۄ3`Ӕfg/J9CZizt1{B} 3cwgz#%Pon\wOu*Cˁ{fiYxUHw`\E%xLNy<[ɲsu˖ՑW$,;m7]D}qg}̭V29=ټlS_dg#Q`l ,Q-H-+W=|J&%I< ☜W ^˫ࣅ1J[Iy_TE|J& xr /_ tX)4}RJw l#<{D5X(ѱ#K$@kp"1,loUnb!Kk9 B9d,184I1[ $>0jl8羞ky몓rc< ~xȄ^58+%WצeY(p.N+Z{k5+Un=[t8Q>-1 ѹT$$NM #@eVI)CiH)WşD킌6aWۤ_ne5Qy=r7~5v\UNts3.o^e|NoP΀Թ9+' kNik)HnX?q؛Ԑ_WE9>Z)^oذ7ܤ5%dnA[Ka"GBZUd[|+.bu6D*m SNe)A0^zle:3Z֡3VrF) *ȹQ mHLtwz {_CsEo: bȣ^z'Ϲk|;Z`Ȳ=eM婜xDȱg?P,|zIlBvꮏpy7]B T@#fx*)+)5LPiNzN#z0MƦm˱744*5֊- C,&h"F\%Q20q0yٰF')&jkrI{APpDW$Fp䂑Ɵ vYdހMuvtt3-yAk rP)Jamog)Ȓr"4f(y4B /܂ B&7;aǤ% "b@wa>kl8YW>6PK् KڣK ?Vd9,jDtʱlJjYI\*7ӏLdN&QQZE\dfI%0),|bȉ'ZAxkVO3~g&gP䜓K!"F)%`{Hg "qMl1dk2c};AF@]b"@tK&Q'K ̽2)8+NXRch-"4rL=kDe>0.j7r(DNZ̼Y!(m:<${XgQȶ!#AǣrF;VvEԤ̋}^uZHWCBlH&Ex6(1s;!DDZ*]2+T!{xtJcWe;r^: lХ~C9I $6 CQAK#KGB<쩵f9a?PpXU$ d.B "Bʓsr,A輩l1}25Bprf[яH%)+o`xhNVպzD zVR $ΗhG\*R(Fa9W,@Eg )PKU8=c/wW~ d~ >>⹑TU{)r%Ń,(Ek&ʤMٖGMu4mH>'N!p&#l9 Gf3ZE`6:/qhYFiC05$M2ɣ<pYHr1xΣqKEP&jۧPjRnLJ Ͽ\/e3w¸Kc 'Jl[ygpQ 1It+O`.}iTatF翏f]fOjS&L7jrt-LdrKM x<x\eO'LEm;3_*ʾ}Z6xt22 v}7iD޾jbJpCgc:+>̧ΕM[w&͋>_'O^?Ox V Z^de[쒡(ϗغ,[=FF3eRHl?e`YrB}vO7LNsWnu6u}{ZqHXm0I)C~;IwRRM4eq'v>9|<~ͻϏwǯ~<^ѩZs%oך!Hv¯ 0?o5MuMͺ6Z79oүkrk]>jovdHO^ӛ(5h}S.g6 O4 ֖ˍO߮ _xjb Alمߖ۸ԥejYǸQG%Nݣd%s*sIOxBphRY&t(7ϝf9&8*Jnqɼ= 1ӞK*I6?K\AC.um\Q9+2#>i/v+hosMx.U0i4tg $[Cגw 4R.%^t C82Z'Ҵ?YPUsYP+CJ\$ɕ &NNƜ9>Tc B>2ݚIøT=hҒ WQUK>29L#zf-$VC 9koUAhRzF|k-6€{r5FO_;?a-k ^[+Pwt=e~GFS7<\!}˨lȦ)"UVz_$\eo=}Kѷoe:۽>Tz~ՊAg搒5B'ҴOݣk5eQqLр' ̥Lfne-Γ[`Q esL9뼓1s4YU$&mA]uOfVaYrVeQ#ٔƭ:dHPfUԥ`h>eo6f}96pQg')9")&+.s ̍t9aY lE*xU(ZG45F f{&> .$ty5=a ͗Z)+#4%+CeUW$ \v[Y$3k-Y8c,G1tXTrŹ2nf{e!I&_E&K6jͅfƧ]r$+h]6\;ma ص.ϓ ;t2dUc9z.Ėya)z'6MIZf%m4z~bEʩ LJuMB-Y^̾1?նD?>7>gy:Ӕ1|<&pDu^xߙxSg׍}-A:<輣ſd-qS?חOmRϻIjZS|INIANh4k'}nUI$JQPzyy4/Ѡ9&_UR 9rIO0MLMR>Id^`fXvU7K>M)M9fNW:O;߿y 1* H*gTZ:^ژv-s*#X++UAGD A2<y-9w=q|vg<E\]>ږW^9ɠ*)e;]9rUT"N;ˡٝvv֛oRxgUei*ShL%wPԘ҉5wHdЩِĀ)h9pjݘn|sUN<'ky)* we]~2N; ۝}cVVfb?x@6h8ԧ]L/?_&/KHqs7v髟 - /wq>7ї_whθuQ.ϧBI3Y/Re.e1Gn7/&^䆾WeYLmD `4{wzubNZk']1IyuWX_WC6sT[_^悻3M%R*XqnQm-Y(J$Hya[,>O!2ڬ?4e3)i኷,#zۻxfO"ph 5Zem=ב{ ZSztĿ^NS1XCjJ}F;I\><b싫osT~o"{s<m X!蠸gJ'$ʀ'Zʉe6DER\m7$9ïB̥|b B쮰ȣMIJc?ĦcQ"`fꪨȈ?M(SZ*۔U֧*U+eil/{y=ފ2*Rnj&)T0nJk}k5x " cks-(iR6j]^e9EVOE6u892_Rw+Tm %e76iRLj1 pT|6Ʈfhnh4B%$g0pf:r˲z8Y !s6l+Y&UB%zEdGW!ʧ]4t4Fթ|/d=;[_:!G/D 0V#6hOv~zJ~"TjOoer#X /ψ1'B&4O;#bŸ.=n iTE8USI{k%hH%%Ɋ] sr=ʜ%xnku}d Sh]RwLMZ%1įZ)|!4fmDiivJc $ԓŢKF%:m1ST:8S׾"FnJ3HK+ ]+(ўH׆RDȗBi "[ݫ N6h 8 u|ԡn#Xe~R7P&ܪʃZ] Eɳǚ[]J3q%Y'дYe}]Q=s%n`T^!LG 둔YlXkYCQ  <ٻVLt&J3EPOilG;!nVCV-42 bf2b@d2!h\F)EB< A2&d]g(M? DB˒02inB AvEmie5͐jPo^]rEAv6|kB C9BBѦEGx&T4v#]7cEfJWuk#>Pd97C3a?ĥ7SL3H ̚` U7T r ֑t.V!joSCAV(J8 s$`Ge^ 2#}৆^앶ebE ]Vkz[d%A/[.4F^n!GAWHH&9Z˨DpPZMil, HPҰ*4]AՊXZ IBufp[`"/Xv׊~)4cMDr|T1&bb󢐴pbB6!dUs7þ.*}t4+&{H\!iR|tLE\-(ǍŗX~AyjN҈HDNW2^i C.knvWݰ+q'K =`Ye>bM=A!%xB>hAjy1i"!%:]%|̡LBP=n;H H jcJI{2֐h NmB[viX< (tg@ @A&e#kvpqAsڏl: EDiv%&HTT@PqkUUaRB]$1dlDH!(VDUb|؀C L4C:iˢ9i$wQViżb$YZ[5"hЭID3 .-zLa7M;-,zwm5EZSRP'ՓFCkׄ bL Lysa#·TfĞT2j7 JȀ-ɡmM1WdsC<܈ܢzA,WmtAAtP%(H :Ш3 Ƞ rc-iÖ+QB|6 IdcҠNn B`|?$o^^ѭbp0) ,*1"&7cQGQ1a`b"ZqO+IUE :ƀ6mӺ`faƕ=&zPIQ{/Aҋ21KCpR 9OٜoY%sP׮BT4~ToZajl[F88mED + (|t* ((-lZi䋞P!V&EυiDF:̈rG ljrs}6gSMhĪj4D2 J.ȎYxY 5I8TAH2bN %:)jI-0 YQ Aygjp!Gf-uWkqbq [!)AVuPY *h.y|ӓbO[-w-\x jfy=ܹY[:8k{뮿DkoN[vWӴ|pq wWpoy͖b VKɊzLmb_p} \Ar( /ó!`'" \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pWO \E i"_pE{b+2wmZ\V1p3WZZW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pvR\i~9B/+X3~"*" \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pWO \A+zA/<Wk~FW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1p5)p5q+ջ_~ӊ^ q\ GKsY QHW0]ώ>-V6/>tA>3?]뫣uyt:O\EǭKYz?򡕏N+bn&~3Yi"h gQ:8wnޅ[jj'r|uqEW`]R z %GˏSm3 7}9&-|bȃI P} 2TF|1v^.ۏu fcz4׏c\&eeb{cVuޚ#-}T'km ]/ -4}0EeRDSцԷ^6{o^~@/E%R v%R\C5՜H9{%&ֲ_ǥԣb#7yvwÄ륫bVw%]xL~+7?ou9: QEoߐ:nzߞ-r{T!< ň} ,e2Gf.hoqx(pSgJx7FA^$ $mt7:5jtfֽyk1 MXO)n>[=+mjʔ_,Sb{ \_ uoIV~K;v[G)嵛9볏_MQC4_es^C ^ܓ *pVZ*eۜ#NgK56vO5qIN|2ӵG-:ڔ0:u=.PÑ䅻3ur׷W[DJJ*JݥG]7fsU_)in_Y4昏ÞbOO =ܽ1J &/aa_W#aGLY4oTAH}Ls 0OT߲~Q?T?WZ[w{2͆KRQJ DӛV5d2Zxv]TYӞk͹wɵJJUBB<ί͏A dz|XnXOϷǍOT̝~y_}R*<if 麗`I-n]sV8J+r"['L MBkCpw@Mj]$"/9÷ ѥZlVԥkWx1*B]^-TTrpJ' S6a`̜;j嚟S0(7 +:\Bɗ{:᳭ o> q 75%~f-t1 cꥷDf.<^E╢P6x <ChE!VCI);[WִqCSR4VD3I!1Mg4pΑŏziOU6#yڛc QQFT7Qw4r}5ǎu#wSCu/Q4ue7˷aDzi NxsקrQ>> +i'kʹ]$U| 5^3CPRh# o[qWO,lS'gdqH 덢-^;kz.w9zYΗW";oe7;yO w1}M?2cᡧT# ٻ綍$%m *}Y^꒍konR@b, I3H) A*<)DzHp=~z+cF7CeelVc6k&Hm\=j8>зudߦ1ۆU޼-ck hŵmf6rTO.|{@|2yˁlD_mb285G>7zE^mYx~CD;@ W`xq8.<+X0b+f4g&OSc."cg YN6}$%Y 4U4fo{}0X:F5SֲnfiJąUxZk[lŻqC 2W%A~i:|'g죇Lmd~ egF$(gHDL4R,]Qew*!e+Mn_ɧ}fXrb̕m @6Rih`԰4˱H hsy !ȩr\{ݎZCۻ}:`;:~j`9t 0h]0Eٟ* x/«+WӻiR܍nls֏e,88#F`q$Ri?5) P5@U[ U%ÊBaUd*Ƕ5҈hn5J)LьeDZB5:lC? T=9CUђ~BMvL#[~q&~| j#@'+ V1ː>몱p+s,K!ܠK|Mb_(o`FN\LoGt/-9W}M|(ڛ8ǟ>x)v^l8xOnq噩[ v|VY׎Q_^Ne[g}xGOS*o~fTSi,k:̈Dŧ{g#(*$B~8s13>&, Ԇ+Q&6װj+{?&joƿP+]j̳Q+PmRE$GEj>kcj$h J*1jK2>Ġ[.fOI' 9 nU3ۭXVHoRFDz+|&$Hj%_Rѿ^VGu:ő%UHlVuJv+m;-0r9UPy4&砧BJ;Bd`Au.v!ddiأb7)8MzH Ő%p.Ra۵t6]y( ",| Jo轘1SP-R a9XuIh kY-\}"۽6K%c< 8ni>^][EsT FQiU |-\")UǺXz<9TrkC=_KN (,A{ ~r%Lr3T_QW>vzjRi4oz[+VkBdsvu63ѕXgLɺ͝i:F0;g\g`0&8(ɰh6TU:t5CMAXK+Z3ޱErGr@}w]m [˛<!l/VX·m+I֌\p~_pތZ.KiFe`f%nÃ4%-\B/șWH=պ7ߕe1W `O6&d)V;+].Ź[L|Z{h5 =S  F9'8m1ӳW65ss?٭ >/qO܈M:oȇ^3ր/M&ͳsBl{IRǢmniOOFs5'I3ey{UOVb\+kH?/++@S!֨G`]}րtMD;vrriFVP/Dd\|̽_y3w`0E ^ш|EEqfw޷ogmϦ/uMwz ):LF1C*2Jg,ö:}ɋjmT7œ. jeQz:z~Lz2\#EOiYc,Pm1֍n̥uc.nL9 UQ&U F<3S&1fq#sg&)8خ|zZn8SqUVd>sԋuN{o\8s*Lk囙:|kP1LZ`,\;*-/+Bx[vpÈ?GZ'roW682n*nlt5F`xf[H{u(趯p19]J4|_644eNNb)2.x;"32B4E*L D:eBȄq Kd)I bBr즶=m2(6Bm0SvD3s[ !>djoY1 Ł wY>&>y,8QDF"Q$&D*9+-O`&|qQ9m_9_jl\8Bm8ߋ汓l;ͲU;1p.cn0g3$*tB8p0MSDm2Tۘy8{i+t&S 1?\i)XWBCkzhxl4cw sjDh%Mdc _aY Čh!&m`m<{@5;@eۊUukȒ[].Y.h7B[PDP$Pg}axTlUp@p,h7RHF[ݿԓɻr 9#XdȼTVɠ]2*d?t f"Of"n L_3x~#+jK48^$A%df]2$CMST+>簘-`4F|'fbBBM6_)SD𢯐Y(wWN9A;1њOxW_ɲ=#YZsb ۧRAIz)G8i`[P min ՔxK\($:}kh7'Z1U8$2 LUnKDA]>>mAc/`c.PrNz]KdVCf(tިF:R NRdxB19Q,P̌S#Pufr:(c894v)TZoBBNyַNve=>ۧrބۦTz%.q+,_gCR$01]D أ\LnS Ɠ*DUհbYFTu;}}/@=Sٿ= V޾#Q9XXj9dBbT'JV+᳆Ib*Lld 膤|L² PO ] gc095#kN͌%xL9fjʚ(™Igl?%3C^*E8Y 2Q5m*2jj@ہ5P-a?̪kP}3Oc fQYACU'1 I1+FlF"УՋ+{,lnP{/cXO^OdtOFQ|*Z#?诐;td M~8HӄvNS>Jo͔~\d# WwŽ=#>bFʩVˊ``l&fH/vS<~}') 6 5Є~*"pwhekX1]yiL.˄$̅$$+c~>w?ՁAH3bK^-~~X?&")r))U?7J%]ckЀL9t/e2X)R/ tg |l0SMx&4-X" * p[F ͌*7z#_W 'BKf |MiXSO{*7lـm<>k]Ez2L_Sg@dDq֝~gh` DZ}z/Ad7/7DžƁ>eٝ3ooz0"P[MWaD{d4&3_]ZץWbcnc]×t71M)wx[wܫX\\jKmٻq,W x'V? 0h 0Xđr.U~);iYr(;3*'&yx;7xu}<aFn{1`lCeݗRd˻vS)(<#p#Hľ ܃EE)0z&zcV!>+uo_izC)Jyq]"-SqE>u/Gq@=e ?ʭwwÁjooZ - -X OѨ(`W$ATz/z|t5 o\ʘn`?xa F{O1sқ} ANa?y"0>A a YNh_șB-b\͉mz:v<۵QĈ㜧һ=St߄+F@1irN\!qfߺjaZ]kZ~&)."]FJ:X\g@B+/_#,jaѷ?;> /X;u$Bv ʰYZRm ci'2)QLBqTs89KeWef{58b 6~*z\\Ԩ=1Ԯ2 4bV>Q̻Zꈣ(]fӱ: zly mSh~fLx-_8)v +koU`ITT,/riwыג*[C14#");ǃ/k٘7Boi"p XK\Dr,P¶1D™R.H3yQJSmU1螳7?AEd 9dQy[PFx;XrcZj/TD $/!}=!g)9 BV+(Lp21B[0zT;>jfmZD !R./ʫΛq"#wל_̆Vn n!O 7!opܬS*2Ź"xZy"Iتy`8CL&ǵ?1^W0sPhVsGO^$!Opf[".&+ T7Hy `]&COg j'~.]V^D2uA@aB,`)mw`WzSB+K&ZD 4rPPk, ÒV]-d,a^~((._L" R \h߳g(RcYhΓAE!2MЁ p*rЛ=X:~"ˤ 058E L$Xmj[u|8cӌ(g~[y0t/ge'%%-G̫Ђ3HYdڛrH:C(xB}C'CY߯/qX#vEq<5$d1WO~RKSL`F32)/NDjo  z ΡaDf\$ѓ1@84Z[ xUVt0{g" EP[2뫱c`U;*w{׀b\Q) 0D2 A͆hI5Js b@ CxPzǞYPbZ%Z\>V7i6V\9WܠfD+=;7"|(H rP&lE2ܕ2-ea&ɉ Ȼɓ9E81vSEiPPض٪-7V `XHtyEx9ۥګr)+nKJn7"5{^I|Z[}?q׷|WO%U7PLdF3b@ 1HdivՙdXMo.Cvf|^ 1ͳRH U2RҼ@DPPMh:xc\j kTFfn8yF0̒$5' ,c@pAUҢ hƪbLV.O]y玨o As!4{kPRR1U$тZFQ-\le\vԏF9BTgC ?c Ǟ6Gg`?N~>ɀlZ9Uq2pK4RwY"A  7F47xr|ҖSdoa 'Gw/㉻:z &tܝM f́ڠ!s$9c76skbl @oxAN6G3P J42C*jjKrtu@C}GD%z.ʫ tug7$f׸:`#zHsx8ɐ;#DݱQJFk\&9Ě~ܒx`oK$bh\1ƯHSpp 3=gfkhWX'zPͰYOiU[xRTh.D\,^̂ zcS.{_FBtޗǿD#}TLkJ0KN]C!'yfG."$j@T$yc (XO)cKVR{.d5'ywv?6Jv2T h1)5:Gf@); JUg;]2akc3}JTՑ<]>6]dz/n 1%2/ zp@F4ʢ@$[Z"XfՋ]J2snwڎ3נN.VbxrR+W8%e7^Z>b:`歞}Qx!f[Y)d?OG^:RQx%fwr,Ku'O=.fT vq=xV\3CRS29u&)tfUZV3i rL{e8c%0ºyN,kkd!LYfe6NߍM%;›/nozvF')yM* c)vʅؐTz_Xq)E@tcU-Kw) 1%吭{=n8s\!Gwkt¾"@rʵwto2^y7̖+r'ԃ;#̎99]9/ehpYB!aQ/)gxȈW/_|Bb=ҁ~ю"6ۅǚ26~el}JqԓDO^4M@L=',*~H+o>fXH"0`Zp2Q 2ϼHZ1|ϼބX76hl;Ub" *bZޥ>aӳ9Uؽ3رs|Yk穾 ^AY dUvtTӋ!>0`$IhdG(o__ٙyl ȼ &;/ɲ[SᄂΖԲ(֓;۲I,VB&<Z)&q+&6kgG+Ow߆ BK[ۘ0!l_8yΦI>ZVKZHHufKmp$NF/dsVN(*_WuXJN%  ˥3刟VpPYn0JXFy2{0i>"yI-˄$[]H)c 5rb)-WG-0FXNqڬCԩ AbΛsЙO19̗n|LHcc>6 'ГHg$nx@&71PA^m-@;MT4iNp}qD'P>^S\jYRxG"+Ή$Y̤ESZQ"!C+ER@uyWI1Ԧ[0a__ IOPvM<'q*(t|i]K'?4xD^׿}s  (]& p]WC] h)sZl;D\~Zg"}!5ϲSLbW)ke7OVˋ{3`]ëثnu4evs֔^: 1dפ/B\uHo®k?u*$RkSv9]cpG]S`ZF`z~ R(R]3!!E J/IJ4Rp ¦uNq5p:N/'ndlZ3g!_/u] ^S=K#/'LO mCWsc$:V3]SMX#uNV`)SBAcJDpw]giOa v(xq?`D㪃Qf"ϰ 3X L?7I?F1|Z=1Nno ratQW۾˘[م㼒ՠW.)䗗eYa^6)iʡ:<8907.{h*=& )&渪O4Rpe4*P+(蝁8vgǰ^8؆|wqLu͝Qp hn43Sx+~S>) i3QU?^Q1/qAs)vR]Wr+9sDWOczj-ECvBs\[d‡yRI B}kd8ZN4VuzBIf9*% (!atcGYx`E72QZbq"jó!`hN:  rԶרA@}$ԆKgrxö5lQe2SG ]PQmWǩ٦0Bބ_Ψ.5Xb2$\!Yu3%gi7|SJ+nxAc潮 yHJv8Cak~ Ǖgt/k &g̔;+RЌVο@Ao:XQѴ7wqM^?> fb krNG͹gj&H?IOnb<&كֳ,m9&fBsI_oiC HURXjFyI>Zr|Zr@b|.B8(Q_mIQg[Y'XPA/xrP*Dڿ L{YrZ?^n{9RJG*!ɮ0JBC2"c-  ±mcƄ7ͷuaˉ<^5J etFzpu)H uڥԜw]?tA~BAE=weU:ƌuK9ГD@~1>|q u1@l|v DǺ5 Dt޼z%@ (H(5F5JM>F(!EΙHe> t"8I(A2ΖPڱ>DZB;0JP:NhU _頧J3jv1[Fۮ(Ik.Nuԫ:2oudp88:== |$A1z_3[d^*pnU!8C2X?@,wKVBBNzSp?ɂVGʛ&u{$$$i u/ai[/sU? P0N&a?夸ӵpT$T rϔy?{whew2 ӟXʀ_fUxŏ"]> ٢-X@B?{tgPde{ %-FJdTX>$JVFхoRHP{ [` /rS2 Ȇ9mVDէ^r. V ɃB7_ˠhY|Յͼr3_VX~%?KN`>ӊ;V>f:@>lhQ[!V}bQmnwb6" ?4 5hwXr:',*OlTB (1D| Rj3ߧ) ??h}ĴW-wdaw6\U6Tõz49:&qipZ 뻘.T8ZJ=7EZ%.U 34]L qL<'8乥.o+f3돻; )Xșz^GZqխhA &$ʌto[3!5uI9)+ ΐr n6)匎ޮŔ!A ey76j+V9ל*Z&%[WA[Ѽ WlkE"bP$.t.bjɹ,d6\6e=Sc4/ 6. Ґ3CiGÇJ1tQdq0ڕ͂(;{" lH v,7:UIwǾ6#Q hbM{(mKψqgO`D(uϡbuކ[ĆZ;19O%=jyѪ'4 ,Du 8w yl#=jc≠r~֔+792Ȕ/„)=)1=wc#TQ ձ@HҢ \_M!{|->3mޓJ[QVk8pkm %c6qK n`Bc7JWaddsźS=:=JIꧻ:={_J;npaA#|/QSH IʉE4uFjc3G$}aJWVѼS:qtG!{CJ˱DIhHE9 ?˙."`^z,t#L0k`I%Fibͺ@cf>e[g+Fs{dt #p'?ɢ$[gق2m$*|74`N Hy|egp%hALSOT+/TJQt}["0QR9 C>"jwNSy^1,0yڷOˆM'# "|0^?R%X-6 bmUb+vǀZL\#,!o7.+%Xxjtofh۬!W]8-O@];l/,OP:(yq2$L*OZ_6+V4kz^rq\|Ԓe#ysxMAXv)'a w`!5gTixsmR}0*颷VvԎ4ʔN>'RFio~TwG՘K{_h:!g6tdV>EXww[JwYAe2 ^x|NR͊,Y(fy%*np4'ohtIpڀΖ| 4?$Jc` #!a, }BC*%fnjG^.Yn%27M@s>TPz50m,I2G;S~17J9~F,&i˳pD1 IBxn&D%4|mYuN(ބ$TʄCǪd0mѴ=݉+84d1hfa>V/SsPE= $0&B@@x`}kW6:DIu^ڗ|Kv';: `TVj+E(0L!f!TKu BB@ach8룮v.fi*+OYɼ;ٝ*ldRDI(A Fɏ$\a) nYRk)?z#o2 7Xݩo盯üe,VCǯ(_r ,Dg8 !"xD2F9*8*TBTJűO#+dC[(Eŧ(QH`(.*(PHdPЄ4<0 {`_1 u,R\ mMpt00q+$0Ho?[6ngwsb|nm4 Eg<~>!I4 ίUS0͍ Fٗ@Y|-.e|2JSsc%=D{KBbHjrȔLeRJRU!x C4Xr#zX@ݕw 궐jv zy?( M@"V) RbASkO_^2ŽEbxAK`PyNWhmG0t8ڼmM/aK΁ۉjt+TAr&7Ѧ~r5Z wwoU٠O}^jK vRF.jADfz2;MBa$%s`ybW:fXb:VU{z +NiF_wI)HLاdvP.ݒ9q{wkXJ f2k8bF sbEwI*Ct\LV;OSi" %kSvg1Jd7ӑ!P@qlb9Z 8f+P4ndLzJ(0 8 sN LYjMᄴ w!RK;-5(Ar`v2 6wIQY\-1t \k4uٯJ O/`kVo˦TB_XLEؾ:|_f抟c?ߑ?5!Ӷ+F~YћR[~Vすssp88G^sd Hj*1@46n<?X K4wlӦz6hVM"98oYtqD۟wJSy|=)5>B`J~ ab!xG55UA!y:5G=r$ ~~_{ҩ7@th'#aklXǜ‘:_OPb|oAy#XD|KLN"edgrr!$4pJ '= ײjM g&A_Z3+2J|@1&:V˒tiYё 9C(u %+ܢ %7Y "Uyf =D|ZA#k/e =~U2B8G0;.x"!bT^xӺh`%VhWu=-#@W}wj3hT(( zlwֱUI _P0JaJera{XKwZs5E4͊/+^ 81x8\=~Nod+#!ORw_puL]We<;hG*h 4uBІ}}~JFMwwC}>[媋 xdlOjfǬ[)f"( eBa+n>8?}N qޛ3`*3%! t-:i"r~FHf-hd)oE</M1>;D4B?Z}y-Jx0+cwumj1Tb6,=fSp_,4OaW)cg8Z`i'ο?`ox4,u6`R}exra;|H<\$Ik̬wbYaEh<2 L!)?-{e>}ub- ȆM$d+f@o]/:?ZQIvҤ)R8fWvnoJjgn틦ҜRj㊈C5WAP3=>kJ^ .4v 0ÎHm OǎMrLq)S1ר;9Ek޸p5Zג#4FB re?YO?Dv=+AC[8?E$2srV_W8۳\;󸻛a^咃p_|@w/n],5:[A0Hy`$,5M1YTL&dZ /Ƿ%~ r~~345?ᓴFioP] zj,Ncc[ڻaݥ<[)-C_ʊ+ZBSuX3*$xXXXTWؕ,F>&=\x`[.Z{հ,gSh?#gQ>?nK#hD=./b+ be)iz0z4: ЉjIs4C_2SԖSo:ƜwQ;J ζ( !MAY3U(7y:/7sOX`ѸCXFfvU,CJ)̕p&.v_ Hotq=Jj#ra 27b&Y6i$ihtΏ|qe.*eGs䃣pK^*N=?r`44^/c* PNf3b\ DaAD#c|l]#oV_jn5-Bx YM5BKtg¼cPeh4 t$1c BPK5st_c*!pYX]L!X͘{:dw>}^D[7WYi] XQ .-;żBJ ϯvn§Rd~B7ZG@~&=hTQx/yLYlQ2@VR6NPy n`tXc%Wi2M}%B[+3)zdGyUȥC=Y܌0+drU+7F&PN 0H) ,CRi@}wjU׃8W._0Hegi7vTtt j}PlL_sSj^:TI;9*^,nc5#%㓃v\:E@_ ReՄEiIqҢ.ň}zy3#4¬@kW'=L@ :ѲN}pT_oae%"$yNO\4w8% H18r[ NV)nOpPcaT?2Vr@AB9bk eH`RX͓N_7ײ㟲-V# AO8^Zg7[E"wpDi&D*iw4 #CgVXR/Q$`[hb'7EvfMb6&`>D~QEvJpzMsApRzSwޤFNou4*~l5ϒ~+hXU+$~/0?ete6`|vo5CV T>րwXl";,bhx6i 8g:* F9biڿ~>)BM 2fȩ$3u+%JVA㍇YȴݤioU(b]aH{Iybуw6AYx֞ |Oo`bV^D_VwO]Y?\^0sF4bt%ϬUxxrApFNLG3TnaȻE!ɍrj[[DfDV{ OFl,m5T`7Bc{ m9kN_A2.~~V?h&7q]ű`>&O8;j5:0؉4ze3. -:#-Jv8}6{5Ec=IO||~>}a#BnoӵaKw#1>a)ֱ4<f%bF,`rQNv46x>Z14wo6+(Sojs^I9wwC =,foz"tԦdM Qb;&<Ԝx^F0%2V3)gk;ND*ލa!sw|˄jTfY7 pM pc3d+T/f&r!ߢ2s"]zRǮ 1 f@: UagObE0J1.2̜ 6~´K9A\{k4 kuhN6&<]5VђJ;Aқ}xg?P20F/J݁O1.b'9uHp5eJykC*3P1ZPAeMpVr#Mt!smߟ`*o٠g"+aF1|* NEh"G /F?H;#SF ٟ,RaeP}~܍*_OgjSʩ?CSiIypVIvQ,Ͽ5[)p@6}4 X*fGgqW6#>le3_-:o"6/=aww? 5[?Z% Q;cjcUE#3*ӡvgp .x%wqAౡ:lAUDqomm\sWQn+B"`sB ^Rt\.ۏrB:Ƴ쳽j& P .A.WCRؤH GؠeyƂE_:Z( (n&m c^4FCbN|ǧFTU51P/ `e4h%\=~t3KFX[@V{a@TP^$9/*Fał;z8_K)-%psEip1voź65dwbayMg*Z9?IȽfRZ\5S04zsښA{db8ڶ2I~n[>~.g,F0OO?Jz[G;@I=E때]Ntb2{]_t]uI&U<~ѧr;q~u}=KkmX^b ] pp *ӖDe;o5IQ#r(59=$=[tUuuUuU};x(ByPN09\dM̠t"<4ҵ,0_*hwn/_ԅSpԳ6ߨRVՇe/hKx3 j5\Gke5a\Zöx1'ĔOmV!xg67J31Gc"8i>2/ILb]Al][xg^tqkxdB>_s(R;N&spm>!aNl:5O$`k#@2D4וƄUIpj2l[žڍ)^iVL|! ;oc =oAYJX%vE"Cԑv,^:u}e+0k*aoH` {Q&X"tͷSl")bGRd 6ȲRw’YWl^8Y _p)cl> 7 ihّ,XhT h*B EO8 ݮ=<RJV}٧G2ScGx!cl,O:pMμ;}ҙ8@wλК3CAv UuI?U4`m |gU5[q/J7VY;x eԲþr쿬7\4+gW2+z̀TttBV9ZQ"Gs\UV^;n!C]l-Y\}߂',qcޗ#*&z_:P ^V u^& }V*hx(u K(dHBOH1;.>xFSY=w83n)}K뭰(Bޅ'9Vw~si8;xǂ ڃ9w5O+piv>䫬PsWUjyAey;hӂwvJ«܃]':H<:CK$#jLxfGK7t{ Rc~$+v&P/YP/}uUXYng\Dcpk OS>Thb%/ʡ9=nn%4Y!;3w 5WL>_.|ó7/j= >r|| O>G_x>\9uq X6 ڑ8l[]%v(j \ΐxfn:k*'>+D* *]=ˮ1U'ҥ>abqt̋j t!\̲*:7Mpo!x_-kSsSTiW؎~`{[ '*4&H:7 , L鷊}W9#ivǪY;xxn,Y9y8qޕ7}\",*c*TZ \<g]& `L9/N-#]H&Qp c CE† A23~SJ2J4JL+VH2+̖|Kfc6`V1WA83&8ᤵE+lhŐ JB#QҜR QJn,5W\!>RDL]5g3VjSч{j?C&( FREAGKbK_rEa}9J :l̯_/.ָH{E̫_NpeUXrjЄJl҄ғ7/c*Y?Q>H~X˖,nݻVBfFYƘg,JV+(c}ݓsTC: !:󬤆Z Kxk:I6 0%j"+.DHƐ!XP̄6^3܉ϳl ]T.#x&S6sh1߫}m!eǍB9NT=le΀Ft$;,\'j x fu$9T<0Akdoᝅ;@'(;Ny{,A2ex%d/$wqy **NULuqn %68f(+my9U \aFG|fu͝XZ1^q_0GvXMGS,yN d5}Qw4~U~]-ؾXqt48@t+9.O߼Vr77{&` \*Et=`aFI<\!52uJZ˞o]o,a]x_KS)}Z s (pnatz)+WqAH}T4\+U&1V| ! aVn^Fz8^hgnƉm'z=z77ϳC, K~9:9&2}>˸v'ucg&OHsx.^c8axvO, j2nB6//1Z혣!Ѻ^X1QFh lk!u U <;a*/96R`_.1at5?<ߞ(cD]KK゛IGgoD^(>][[O ,ڸѕUT`;*ې5W?GxfÝx18ߋ_Aǰb:K9z]c.6$yIֽs#ɻs3'5m"x@}y3]8oOR}KjJͣasE#ŵ2]2JBXG.c}3"露?/ʏ.9_?z.KO a55}p +NT6Ƌ?oo/^_`4Vk>Nj:R>Ҋd.R >ni''o'9RD> g#CDR,qct{ab_mhk $mqۙ7 g.byݮ)VFE}BN8pqη9W4oH*ϰc2VT1Yi,u(%N9ŚHxj "w4Z J<2Zt)\b pm+L&Wop˫KA@%@DKn*-89! 8F;aA-عDyd` 2D/B0320[;%2*I( h@XL 42G$q։0"10يT*t& f*xK8ن,26\`>D^)\;},X Qt ŜkOCE|'aJ %1=(yA"F*Lq831H !&qnᨨ5kt.[Zu< 5`ZTW6 [;`kH}@t%* WUeI]if0p^1'` n@9*SFe;BbBH DL)-0d3`EpJ!UgRzw_Th ;LJ}q7hi=sT8O}nue@'Y Ò`ZIRv 0("TМ!RZR@ ye C&{Ri ȃshL-`n8l$͎(p8 Hҁ@YH,4Qh0hHqP7~FjIER*K7hKl0:Y!]l UIbYfؿ>R0RmX ^?m IcG* du;]G1mA#O !JJps>r"ٻFW $5 e`yX9A h|l:KnվM3^.ȏ" ,$벶DK]H k{{R M#q)E+aU<3) i0rT4*}IbSdguIh [+I/f;RH$c[<9[ k;5#)jgb>p VGII)BRB]!PdɎYwZS09qB'z19i)]p7%w'S@`݅#`iZK\iZLVk1Yd5k1HKNq@{Maa/7ciEc}w9y=^~;Ѽǚuydr>><8An?K_horXL5NjL+`g.K\o7˻ζb?n94#kg{gyƗbpGUQbGL+aXVJRGU*+w^Gl4~?IoLDJJJCSOFcG;jtSPz]i@N~0|*dS PuΑ(5TƌbٹsFDKhtN`mh>ۊK:/ި2}~:iv7hCU` Aq\ޜBiN H:)/MjfJ*a|{~Ξ=z~gC'8S%8ޘ[!prz%!%8TKէ>X=[ad[VMxw7`ֹYu9M@H!Giif Z54v-د͞mH<)`.\wgQ`̧1NCjҸ pv'+@/FpUpX1멋B 7psUlY)郓 R&TP}=[C!5x5\O߿LW Զ@26%Uc3ӆ &d;{ѻq.caf3[FȂPXbhҬ(C2[C;!{cijܾ䂶 1D-HL`ET揮|%h:+ˀ`Vh"$r^n):Kϖ!@t:Zp %k(j{kIw-bpia2^P)JM.NQCZ'J oߜvNܠWJ@c K!{C&uGPT:  V.)L%BQ˒QiuQpu+l)B$b `: fJ|kV5pcz,t W='IHYZ3 26wOZtht&^nCusO=#?~ܷWqt"ŊP"E鎔xW%B v9-fTV0fk-dBUӊ],.g bhkiķJ'^HƻL 5EjOCoް1X2a2ՑQh9k;"l urY_>۹c= ;v_nmrkA`rb:-T]f3o!c"6$ ',cm {W|/?ަ5\d?u.ops:=9 9WjXGNM^ոhl5jf鷛ء .Iv{IQ Emo}ZrTfЈH t(jڞP}`4/bK n ۴W{ 4}seK=*y;1iۑaݱ hHñk$#~Lj$R1YȌ]^=0mm4]%{Knۣ[܋m\aBzJh6J0lEQ!EY6ސC0:So'f4{~Gg ?"Y0nt_Ms!&@ BKrF"YVӳa,(,uhI@Z_>dK ,\;yݐ69?5ssӇJ¾e™Epf: u@b'Yo|¢ XG1Fj4?,8ORXTj۽R=ؠg3Y!eeNh7+6 mMm.:-wI?,Ҡ;{ެGպo\Y,ވ \XVPjVо$1Dn YTІzHzg _:+.H} s I&ٷ]&0m`QiVnKY`ЛNl 91ؠKo# ?G؉V }{m pϤ]ޘ.PP8|zAT~YryA>)4.9{| )b_!5q͏Znsjnzt.kO~zS2ּxLc@q{7RyѮm{ XX_e$8u9㗯{kq\ZtJ|;~R%7Qcшz]k,I(6$M#?|,Cl!'|8kK|0 .|*-gvDzMr 57jub2UZr+vd@Q}=k9MŐS؂uC!̚yېܚ˝OJ};u\Vbx'fX%G</9gHo+o,=/NR2'N,O*%bɾXu8mzXIWBt%D!-*y@$rE#kЀF94N;]KW~OWI\*%h9dAڐt)Z%Ln"~QoE`Ѫ2=&YY&i~4Y,h֜F "9Ι.؏@%5M@FHm(IV黝xmNdvv8}\7s+Rk>%Jh b&1Ѳ:I?"-AAۈM1"C)\rP吥2uVXmi;C\XCPB+ ,x0k# \nd-%S ɡI P&|\x,`3>dBҁ!6:55MD r2*r*"-sF*/!DI)3Щ5ʄWmVt'c`:j/SM' /FX*PY8VJCِA^""+)5Z)'I%8E6J}&mh /!VԂzK66p .ViZaKd4T R- f7=~+5i6c<~#?=?ڛūNs~D@)֑xɿDp0AK|Ed$l mU'nj!yM!1\RX+1.Ԟ;\͏Tͯ}*媬Zʚ媭 <52V=ұ][oǒ+v#n<$޳؇@eӶ"IN~(jDeC)4Kds]]]U_UuYm7Dp΢ {m!Nm#Y۶Um;-e+=q✴XrǠF?%CmxW]D=PZ5ᣱ#i|461V9 QYZPI&rjP0ǀILOʍ1  yMʛ/W_h@K 7+uQϫOIb&B0Mglr*R;$C[:%2L0T;щ}9XkpݧQO@S5PǛjl#LAF5(DFm@VОH2Œrn;HHF1<=W7T7QwwQV!{Q &VM\ۊqeUU\ص R7K\ԊP3d4V!ROG+%lL_y,lplp@qW-aT3R>rIL*04ܴ1JB-V&PTA :bl RI¥j}0P YhE85.ف^ɔ(s– @ק䨁ؑuW)Gm/Tz&zdaUA<AAu1'jM7]xN36*o2Q79.]8緾/?,_~i1bAIЫɫk~mzҡAG_)  oYޞvEsȡя_5?V~zv_Gmlx=ڻؾM/1M+ 9:,RsRӱ=|8ܳ=wvp0DD`ȇx-`DtFW qDD5QG*L!5B ẃ1[EJDLHI:;I3_=yFsG[lk})GJYU#w53*+m6մ)l1Fq/_Oӵu9VHWlxlirL͢CtC}_*ck'#y`3ynKqVx%[Us4y[09O.'3CӿN.ϫeF*~<'b5wmeלּԦ hEy#.sz#^_x>@oUwU껿~Pa~8ar$X= ۟>f*Zyfsd̫O3O?^3dkT'6H4T o3)LW-5g7~*̓~Goۏm[@rÑwlck_W+hQA(u痧^U8<K,LFg QngDWKu TJ&Y%-{HO M쑀3E[՘# ``E 7,٭h \ͨNyw*[y2 %П\J9d'^˜Wojoyu ڦg7C7)d^ck{xgG,GW_nnrt*x+g9X9zp  svrn(Lڟ 'W=o'aj:4(K]9솒tZCc89\g*_^ӴŢubrnEZڻy>qEw@W$~># j3 wPrAqF09`[4.+rUgʬU._ g<Ȑ m|$; |.yFPSa3O~!&&8 :Uş.2iEb .z f|f-3 \;/I^Flxtoxv9uAuŬo L55(8;%M;DZ"T77n[N7;b1O-k>Pr5Cp2:%qH%|HL@P0}uv]=24rؓWe3/ 9"kHg$3{,"Њ* jN;9&Ѡ|&&`cF90b}ܵ&`1l)xT!8f%Ov5K@4Nu *qk‡7_! $R gcV@peرvLMԺLM2m@rd%2;:`<*lYKgNqᣩϕMr$KiOtp҂Q WүEϲ m [8ֲf1:m zPދ5d:g5F(yaK;n0[*a(R|=f.-^kH]i]Qm5!w )ujÖbwMʥ%ITe䔓*XGMV϶c !aZ?#QI̡M ]!FNL+;U ch 0ިWruO!T(jc!T5ŜBh(0jH;oǰZb^n[jh>wm>y"zxI0_KvAluJ774tzNܠq@ |jEbIqM;lbRryH2ضm-xީ-V%&"Z6 )9:VߩEC~Ȭ4ȈH"/Fxb샗ф=͆jt+ Mi66(YUv3+-ѡWt{Dc(K';>w D5`m1MmnE5j2k$%:.f(z4e o{߉*uYXKncٳ)u)GϺ%xy)(F5C)|4/xJ?Z 8B{EX| 骘/ +z,Bf61br7]y:ʠKcpuw3l7cY`eT=+SVt_c6ũ:pK580SJO=jP'baa[ﻢ ;rhD{~0fvfqYpTa] 8 Gڊ~?s۹sJQSEW z}B~Z4Oz q eKtL)"vm7A,(:60 PzsF* \cqv9tƐRŠc#lHOΡ d.5k~sQޘ'g[ʓڬ~ }X {x&1d/H&u?MWH2@.jvv]uiZui~]mtZR:SĔs+o`]8IBRBwXT?5ʎ;A<{yx߭3nm^ݫ.}gȿ\[(-bCH:8y\ zi{>mHٛGDm*wmq~9/6,`xg뽼#GM~=fFH=#V%OsW_*K*rR[+F$ *1K+.wJ|DϬ]1.Mĸ‹/|5 {e뎊yDK9ҋƨVYym'j< SX!eꪂ6BЄYf#<$Pq=57]1Xz\v=Ě)i=|ІmǾ52 bUbe%Ƃ+``;Z+z(Nt. znyV r*.[ѵ( f)hřA<tV[Wa8fMpLB%dwOJt#/`om^`ȭB4 7-*-<⪘-U)dz!E _ޟL\?gYZy>st\e.s7ok{HPf?U:UCoYsѵ}`yg $!O 2zJIh f,l+%>qXtTSf7oy=7f7ACjyr3 <ku:=]rN&f u*SObrJ%:[<"7[2-hY9A+ՒdHN2f]j0nzP6*:FD{~zĊE(©S:j #+~ YQD-s^^{a}sV ~~Ȑ1ƺ/"n]EEت)Zdj\V|}i64Ũ]'ډm>hl}gKxW4r]z;'.ägmM# Ϊ9+'#b"YC MKol/(A'xrtB3ps50l;cD`âks)%LE' pl?sHG7E4>>2A]b@%8F*b@)l, ƈm9->ĴͩT[ Nf70)oSI;6SkN""Kɶ=iմj U1'CU櫁}1)ϮB GHUfwXGlsv: &v8&IlIύc`'μ^Juya_e=XC78 ?!/Tgvn?TG^T. BUW'>-Cd `4`L& rčH,8?VThcpQ8_ k _z/+0kSEiQ6ɖ5MQ')a2kLĸ sz5"STd q1'jS;S7z)xMDD&jw3{= j'D`E*]P+06{FԎ673yvOц؇1؀ω9w*0'jgF'jSjw3S9dr4/*{?=X9 04NU !zV9T9n[W}ʢhgebb ?p,L7<)ޙv.IϪp|OGoKq |\wHޫNaSc |+3On7)L%؎ia[|s5jq>;ˬg/>ޏC7xx&- <h5=%N"h_чOi<־:%\;)c/2u]-?}Ǭc{+8gwG2b_ʼnO~['p8] '-bǝH161][;kй]m7کl۬إTkч7F>xמ~ۘsܵnm%BaQ~bݫwÆM J"xC$bc [Gl˕IeBi% Y^Fo6w`tN7+N?}RK5[PTp`Fp;za3@o W|S3|pql`O;RyŖv'IGA7FIÎjw16UT8^> Gk+Ѷpdϧd?^;AbjI7 yڹ2\Y mN3a n>fWddo{᮲5O1] w{7)-ӿ}\> ntִsusבuh5"ѴHtn6m:-NfXM1GW!Of0L&{@tDp+҈ &튮V ڸHA݂&[kʝ3Bl,oOdWoʔwbQz"k^H-(D!4+ yBLٖRS *nXPkQcFn-!ƱőF@6k@N] q ]IO|Zy@VԀ*YʴɅX&"dѐBRB,"h7r`s/3M}d3@(\lLD퉼WU=W*W㒈iILZ;x&<Ϣ^ ޱ6fDũ򑌉Y搫4f.LZl6ZMЕ%OaJV,# xdfǪL 6Lõfߺ xP3::x77 T\(C^2/|o*;h %FU5F>%TԖ[B/6]9%:S#}ĔB.1/ƒ[$+S,9kS!Ү%в)!,cšt$L6M? ݢttvxz6{5b9: ]HO"n/ Պ `H K *%T60>m"Ԓ1ihZY\".MhIN' S&[MIrcfP&4",? Uy!WA ~=Ya`p.sX\ys]g`d żzRki#x1B`їف"m#ZF<8^lA5&tӧ݃oӪuڶ摟x7LYkӓ)^)i_ ޯN뛓yQY+xu贾E7V=uvejr?.%hk6f幚QsJ|vKBij8n >X̣g|x I럀+(h2!%11%+pIPF6]V糛o߽C=8;ȷgc0.H* av "ozxd1FFs6YeבViɅY{ֶֺ"ᴅ=/w17֘dx %fܲA".\-@ACOs`>_lG0X)Sd9˙ LY ]CbIJ#H?|su^Z.a靹^=14$o@NJm& [=TB -5'I&gbdI@2֒NPz/U@(I 7QgGrFTXRuK`9d'XRe!5SfZ: 5C90z ?}"R%~z[ Zey?V*Tc~کji6*MD+}UgVmjj5u;bNpȬf`>ѭߪ'Z[u=0l{;c{ )yOѢͳ9rVZ/(Q9{z-)!҅Ul {oۢͳ(UpÎ3 Sin/SHvȔ(6Bk[6`VǷܫRX!:ftHZ([ PeRaMbu TȴeZ i`T(% D9k ! )Tf_t oR1#u7E[DT`lX)0uMh/2p4X5g7(O&'WPX3"f[?D>ȿˏiσ ue%~?y{ogCLNkmQf1nx_S]2t$%3+jzdS/&V ,Uæ5S9#$!/\Dw)kڧN{I<JAs5cYci3 i$V()Y%:!5s(kxCkQP9o*-KJPJ$L**MuA;> wlj֩ liw)\z{RrL@'gՖM 7 ڬKI%T MlIDcK\,1Q!F&29& $9G[3k+5= :ǐky!¢ fb"fQx%|ɔdH \1h$%JA޴ ?_bA[|'b 0&8\^?hBYf266Kֻ{qIWZ悧wѻDp98h2AFKM@⺔#ed8E:yui'ibtl%J+'X#/bDƗ,D\\Ƣm9dƿ<H1䙗\IC4|6fiiԦ_9k3Fj9[>0L?dvmͯM%\\y&)$XP>$LZH35xlYm谇HL91{uؔMDXx&qll=SU"6&9]LB6YwPBc 6 X <\lu6|/&PKx* -Y_t̲KLA` ra&N{~w#{E]7;GؖmvmfgYk}G= )2Z ާV5 2&&% BW[gv|y!Qh#ݤ|@FRhwA}:rK=<=*hlT*/~jwK$;.R㣟?SA2h.um-媬^I$gt 6&gU$r}j!UΈ1Ya5h%tRz.Cv Tic 9-eHdW5kJ rQDRXoc6򒿕C0-zkX1i*72'V$[C/ * =:!emD'U QB#5\ ѧZ K'_rlu5y|(/qzӜ+׿0o17}z|ֺJ9F檐oKVEj*?=|c5+j~1MOf}? Q| v9r>}*<$, Q4n`CC0SڭYҳ;v" ^/rsʢɳ _Q &]h֞]ѧbU+G _aFFΓ /bQjA9> j35BcA&dzEΫI#Yy=&r%NASA/dd;"eVE89ĜPX({9tn?:NQqhբCI^Vx89v7+[UFfu2;mjڗ-ʫ9Fb7͚D=̴'!!yLwwT)GuR/d&Y-ABt#~"pV>D E1!IFݥeGνy|Aܛ^ c Rf] 0Jy0!v<fGjwWIFE9IP )jڍԮCYؑ}Ԏ]nP;#VຎHQ}Β2r5BuAkGW~r]8ˣ=o_|:" _,wW CЭBŇH}*lnԎ}iFE'VfGޡ c`mvA5m_ٿNoT?ji~ͷR렶/}Ә\~GMX>] -Z@7bv=٤ڮ?cE?/2?]8ׅk ų՘5..V_ ^p 0ѬGn9h EΚo׏>]~IR."xZqzzwSj,rMgM}yrי ~0I;Q>Vf>~2Ժ=D-puţf4z64*PWCQ4IaR VĨ;@n TG,Z dKf8f,8K7w)6;y},n6n SԊ=Є8yݭ}_\u(=t,b:z4QEG "_ΡV txNVoiznha^ H]-mxGwHg-W ieGWjNjũq__ӋdՋ*dxu ˪Oo%ԘU'HOC`&K( D jB"`kB}~9y?1sC`1ȚJ~p6gϭCvT8Hst,(-j|4Έ8T A2rDzWk'XГ44Lq [?UP0Fy I,NY:F5Yβxm2~u-!45t&t]醄!;٧ib}JXdYb# PVx*8E0)I)<$ˋi^A;td'x˻a:E]A-򠛹yqFtz0T!*R 8IVey".ϼ0ɆwF0O>gV bM^ qxsʹ}-iףsZF1j̹qa n{]H"teaxH٠a 㢅tgzFQ@_j Iov!}/M]'5:sU+W:8,eva]BC1wCIH:-( T#.d&C 2/*|*d0 98iI1A"W $$N'IE{ juN{Ϙ߈t|ܞkVG<ʂyEȴH>̲_ڦIBZԤYȁXnDMsPґ%y&E!-` +,sDcTnmG=]XU~зXSNfY2ς,Jo@򀤕fvCdjmH*f} H`y%ZR=&o7i]rWoM >EUl1%RÂ9΃YsEq -G^X#3Dl֑vɄfS4;LsIvR}.HI ckd׏7hؓ7PoK¯}ӓʰZb@B1*4o!^70L+j7Gg2m$c͘)(:l0BR{!E"1CjDj $cԘ!U߬{]a,q`zt5Hrg' M_C5ӿ4+!9O; ".xYʊht}r3sb+v8 S-Uffoh-^"2{1#M oy|iw,y_BYj=gES`bcnFs ؼ16D]u?nf֧-Ԓ p|Ȏ$Yf]'v9+G@TX:7ѽlhvyCLy1w`QC!H׮m˙Mo^rQZ="Tf[,UV""6bWٸ"bVƼH\aͯ!Gѓx&>S;1;{Q;d%qLO{._scAPL\VZೳ[>8h| A35 PmR*< di ȸFXv%!kƴHÖ]dma .X,/sGΞc[+Nx++m5!3Ӓ ;9,ܷ~{i[~7?wgpIVr\CΏBŢT.Án'yd\a>w402ahn<)[d Za<4w*ҴI@ZIdli/>f< W9EBp3?W&(XXl8d?[7V"MDut QՉ9J}HdmE FcΎ'2Zy}l阩qwjI?(ϲ$j._=3b qa5 ͒ٗ3Ch?yYJ濾 tIѾeߺR)]1h*\+>M, tYAZJ9}* wSVT:3ʖ&y;ɎwTsvjCGMW?zDtQ@^s? G$W|C2[~&m:թMβ_M_/.^Rt &3icEޤeѲaRD ">4.52NϿ&f'N*T%ˈ~TPb [gS&vLgmJg UA5LT,ޘ5zFT Ø@=]Z5lye8zgmxWH&y=~՝tNjs`K%v"F[w}c)-8Qz)7 /˄`E_.CwM_+WZ%pyfߕ.˛^;(1KpIY^&̨܂gqq^.=`E1z)4miJC }Cfڝ5~B5Iimp}s.ɬ៶"OJ4zz'ghvԵ7Xq:^>Zu>zm݆wUڧg劣 4am$?Z+g};Fݫ 1,Vp]r`dA'Xavry\3(L1L@H=aJ老y&qN B&cde;Xs6oF;{gmUV}l6nԆw< xڀL)n]: O^9/{>luj-fv8)=寭I9'onzEtAa CV٦_?~=}eXz$1W33e [6Y[_.͖[tY20}?b8O2}$b@88{"%6%s$V j!DfXhlćYs{r|aZcrROR gR;*dMiQʷA68pF(.a|(ciU َ7Emn|n|n|it 2҆vZ30F;-6[ +c3v_ g,#j V*u>:D V>4El[BrAn0n#fC%c#SU&A2Y"=+۫{rt<TױF5O?"X i~,7z0+|ģ=1{WF) G?WYҞy2B< oK6WbWŃb^c‹qώpxw[!mϋoV q⛌5:G3j:t~!9 !,\ q|OMl~^9ZGI u̳gf [9}Pi' @dJ龜t)a Z[|:`3u?Za,ݞנ%dAaK1O@,Omƕ2YvAT24xV 38 cҏ Z9nцw< '9ؠGsdDŽK'}ZQ"{O;%OЙOn 2xꕆ'Rn~J΍I_JKu3Rs'`I0MGȉRMZ[geL+VX-0\(+cy?{ȭ0/XS2 8s6/ ^gI }-jjI/ 3M++D>i#̭( E(P ݹx{@UJ&V2'4*q$'&Nsn=L9" s S%>B-c^" H27l>_l% ]CL%ǫuZ?_E^AM#"hH]P_T]*UI*:W% xA x.Y ǒ5R@VE5q d FWɣY ~,y_ϧKAZ],A߆t ˭Mhb|1۬[6Zgc59#ɠC4Gހ&2_zTD0r VxAd"`Bz=̺(I ,Iavr$iZN5r{זyȰ-o|0rc*FL%ۨ)xCEe }bFduWR*TM>:a'vȃU +"nX'-2׊kLf3#jr4k6 dNr'!3^$Sn iÞt32]Kj'5PoEb뾝p?f^e!EF*\ESIVe#XHWs*DQp|@R#3"j%&feU"f"Ӧ* g_cA)7s@Ju= |uTRIM3%CSQ-Vn.+{}!-pƨcy.mm^޵nwЯziFkջ .ѡm|~38eRWCc=T>1$+P[&4̰Y0eqΒ<0\|Ph&+"EREޡ8!]P@kbj*X} CÒbs䌛l_Ùc$cNRV>~W@# w% GkKwxa.#h40g,c\o8H*l`En 1 %1:w6=wl̃_EX@rc+1 OquZfK E(ƈeNbW'iCٳ{J<X/w0gP!s} :  VdEVʔ Zj!\<8^daGt{0qWE"3(K4pI3 xV- @]if;̲qgJn_=^TD|;]Aᑥ OgRBiq#Km^ 1ẁ 7)5HwU:Udq` v.MUT.MUTͺ4kv7'ZѲX*EJځ&D1$H4Q&bw 4'[{F>JPOM۳P9`|Mk>x_*蕊mXvdwEO:U7h#K_kTlz}>C < g)-uаHuYKm!زE d`Ec qY87ksa7c:rh'fǸv[~{xI>^aY"e&VgR+f1dJ=I 7#(<Młq 1JVAv\(/&>ȹN^9j%N״`Lc}ym ݕ@*}H@`ljsBuK7ߩhSz5YrW9N0XXVBr+cAWZsqJjGNWf#:M3D'$f$,ᬓZ ]h J@b+tzx*D&fU@ɮs=>'ٜNn1Z2TP.|V>bY$Սr@gR״{WQ2Rk,\r-_Җ?~ TGΪ{ $ RK8ޯ>C˙=n߉ƕw3|q7G=j =<˙Kkun/ĥusݕpm{]ۦt;] qһ} 7Ȏ9-?Tzw3VA{1?\9ƕϦ\-rY էSX_%=y܂Ѝraf&_ NjiD,W Tl6,*{fh6;DȞBB+=tTCzZe yHc99g5܌c':6 {7H8?8'm:&`n_xHHi%?udAU/-Qӓ49-VԽ>֖o98,6b?yr2X4JCؿuO8!3">a2{ӧf\5_h̃A' 0zYзܜ( }vMDtG{0pw/.i^\n2د*siE/e.:aR;pSӞo\lC8*R '(+礯,81dOKSʛB7d[er]9e.oZhFL x%x{\@٥!W,ĮEߏo=_pZK|xCTԍXi42otxCK:M^WUy4qVhַS bw+Z-D5D/'%4a֩jRxu<UՆxL픀N¶X9zL5u+}X(0^2Fʋڽ]ḙ³u7Zr4-5QhP5ir9V&̢L2.3Xl%[{l'EHe=+ -$f3𶖉mh9뢽 Z?lNCWVK ~"7ܩ`c@p{D0^+eE0uɒg 6*vV\tob./@fZgo6v2ӷI Z4&Ay&sn@T+:O.9UΕy|Ǔ" f^BtA>U%-;s1 w?VvB~+uUm]r[WUa2WLq4?g4$g!rOpb.*e4k#yLq%ozзX倾%0 ncj4Z3c^J4|"_,Ҵ:%7%;Q+k-B?.i%i9KOl9sQE Aޛ2^:X!.!&@ lB :,JWN$>nܽ &wy]qhT 79?Ժ!9?ӹ[ p٫{p惝 kBu1]9ndR+N\!\sB\k2=!pDThszR:-4NeZ`d0'LZDP]#)1JŔQN CfFKNc$Yq}'toɐeb2t2X|䓆CP$HruktVHt++xYp`:mkPߴC\DZkPu OOǿLu{^up]+9d~w ] RKPesg`\L]qE)O^R.zH &F 3TP4h$4v|eA~p=ʺѮ4VͼNqwGQ˵YK%`PW(oe )fMϧ"cf@k%v[|mJ N˵X*dEl^vϖcRDN/kݩ6 lľKtZM'\+kO0B@?1ǎ!?>~5Cw>ns&(~CsL'v8vMC8h*)9T;1lO@=u5z&IĀ~>(ө[1I(_ʌX;'yNqma~~v]1zVۓN38,5)MƳQ6׹cEIa贙` tQ L LTDJr)#AH].n/~ k5)C0~5QU$LU UU8VGU;9? CԜe<\Fyr68bv&_Nׅ}5@U vBg_ L"L/f\ȨDsAa B`^…Kt- sZ K wcϿw%4 pN Y9.S rgfcx{1V/EEAe/oάqO+OsKW7Ú&Z^[i69ƈ&ྰ80)$Ag%k=,ſQ/l"a!lcAȳ"er,6Hg L, +3I9QԘ nqRkc~H!Yt%:`0BݲU  d sJE11BJezK<49!TiQž2UYƓr-2ˑETi.bkn aS ^\K{-, RXy`@c/e`5z)|̌Z UEgF=d3_ ^Xn Dwy2x`+5U- >rcyq=k}M,,2^]74 NϪgrGhH$z |'ȇZkDZI+8Yg ʑ-~1#v ۋPQUjT {7K9EQh餘5 .wV'ysa+y5v4@ {LW~ x8,x9x^8ɨT=)]ma Nd(5ͽ?+ƓZNk6K&.\IlZhM2CH3%%2YMEHgG\$Bf ^ %}rlVIӘ;޶yAIiNT0";6+,u3O.L$47fUPw3/_Ľ O|Q5_ԂYK եxЃmP'K|mn܃FmtaI@Tm2>N6{v:ƴ-=`h)';6nd/lxӂw-F^Z !}`Z$p[+thhy?"U#"556iީhVmQFCnF'Nh+OQ['j6]`߱Ws91RߗSp8K2ڎw%&8d9"L~dM-YF2k*U,㌩i3jr )L3L Szy-LiU S;`zi/7vNZ~GK5FjbSm\˧=Of2])Ybf'j]_ugd"ͽގ>ƌl3D>{8iu+t^e\L(FI ed϶ѣURp\P: bjum>gz2;^^ˆ.K?c7 a--vlBl[ݩO8\[`gˆVjnO%8r1_ >LN c#c:{;Q& N4\ # ̕_۷_{Y-CJ̐Ye}+'o'_;Hgg3;Q4ZQ/E[^.ևӟ%`'h.Pt~;}:f__R O/^Ͽ>]ߞM`UMҽtz?|;-ߜT~~.(ljem@~>ߞGhB'?[I rw& Ep>>|(ٛ]7iJR\펞_zS;Mvto Ox[Bds'^ 'srob^{~nATL~O@_@ޔtw^¬V_@MVsI* y ֫^/y9iW:K&uddЎ#dj"9ǔBFoSʥIUdB"͐4IXd jygo@>KR6 q jm;:$']]LɷeqۓsjHݧkR#061>LIEkK1$g sU8۷ 0us*P[0ɂ'걒,f) &~jVg*0ToLvh_cʨÕʇ,%T`R)g*0(O>GUP`Nb5/h%7x@FJB K]TWkk\\VcU3ᲾM&m*`'mL3wo檴 aDwl_Jb`4S7z kWjBns(ד"?D')-EbLD.38igͻ,>?M?JʮE_MU}7/Vh(][?CRm9MwݎzYWvqE+ 8c:Qqw*"g8JQ "T7"i^v<+Y>/,YYc\'/-qykEg,䪟.7V7˨4ܱ0 \^'(m1I?\ beJ4Ʋ1MRE1߶֊Q]@6uGN7ZQ>DG K;~~w {/^seiTdf}U XXDSmF1?|ظp(rہ\o lc$42ʒDD lgWF+LͿٻ޶qW|(ȇmf0o.wp1-5푝E{(y;-q=Xy Ҟ/=JZ l\ActUOR.cTnsJ*\hK{@zIaa+M6ջ9I n!FTۉ0Z0΍UMW<kYFԁZ f4 .(}dXƂKV !0?螃fV>|^2a*;Y+Cm|vW9*¢|h#D^6@ަq= B6Q(X6^b% ; ^X+Q]d1y6gםW_d?GfI@Y'iݲ~I8jmf 8,YRd0lO^̻vMóz9|.0CSkuҍ]aUENp)CSa"}l YZ^*#)%d83iS[dea&Z<| LJkl2LKvg,gh<Ul:HC3!d:aᗈoapiGww`2j4dA6 }:m2mROcAkԁTF"\8D qdž?䏝[y lcyع"KmP/X@ CȞG9C ,3͙22.8E| &NÃK@RVbp:TJSiSNi~l}S=T|7zO9̯PjM@0% $}ʊu 8b0R8`Z@1-PK'E:ZSIyj-W3M9^޾Xpt挠 6 n#i O@!PV?elt8ϝ@\ Nx|F{@ :yET:4 QZݛat"۟Ok8pA/[;>|vӧOWDA VAfW:XĉVF-du״15&2Y!5&dט얚v H7R B# P\FVaLhv Rr22 . >"?NJgBr !E}_ׯ$*NGkWR4 _t ¾. Ađ ,0<ZCG+ ‚2^K.S),/c$'׈WJ4k l\2 .pɆKF.iB M`?k[` _\RI-mduLy\BbZH[״ !./y'md%.b~쩳JJ! P,)JHɨ CeUE%\ƌZNTƌЕAWHw@.ЕdFU"D˂1 LHf,e#HKFHU,}=gNWJhhJZv]U)cCmj3ٷO>8r\U\Xi|;AD.]@@5|c>=}5OᔋXȂ0VbkJ ( VAD Y*jdxO%j/9݌JXH_OqjA@iDoM6arH_2IWJDXJqp39qΊq3!bUadBSfKRĈIB"QL ܄U'f*Ed{vԈ}[3tArJsc/+.֨3uk 4%thp€Bd(pNT:YfDa7&k S`dprvMTm(5c`0LZh\eqb |0Mu6H*1tKDEgt3"1X`nj/Z"`Axෑ;AD !XH ͗pn.J+}}! D$O5 ns.q_5,r10H#!XZa#!ẚi!ʘ""rM1D,"<9tXRP]"!5$щ+Ռ(kCR_S {3\O EhLQ@$#@#aj;4bP0" 1U6R},Aĵ0Sq =|šrk#%S1eIPE$[ݖ̶&z-W&ZkJ]'-5_aeڢ ml e50\/8Q K $n: t/Xu>^I5JqYw/R5VN$%4RT0JjRuR+w l,锁opxE=kUjb\/^ e%ln̴SZZMR'2Rɯ`<`niǕU:۬' jC s0X2е.0ֿ41_My,$ Y$]HnC2S!a`A7ABcbkkX,hVCJul6z5CIhRsC$Tz/A{3p92 ?yf )c6$eI,_\HNж{d^g4"K]oY_fkm>iC3&,wn۔Z?6{_~rgHG>6ή@Z-/0J|R3ղǂh] Eơq,S. bG*[3BKe6*4NisCELL)砚3/^|4mʕV*A#S-҂S\c3+SIR޸ CYu(>[uN3j3ܣy児B옻)C_'Յ `P<WX)#2R.Ukw^ཱ5͹ ЄBƢ?v:$Y'۱XEs&hY\5mk۳.~ d(&hN+2!A FUmpL Ԭ**dv\yj̺8d6hAZi+2Fektvs0A+${*\ @6?;ᗢsA\ZV}91o: (xQDV"A B,W1b'a [STq%!G5ߗ)qN]Kg>r4O1gm$ 'oy**l^mAUnzQ"S4^VQtŭu\srp B7:!؈E#;1u^rEeʆKp LpA9@s"x$^yxt+ R0 \pƕ3+zx6~uAl^' U.(ys\|Rvr!q =a1רZiv{+ES @dpBzU.}պ6(#*D8v?d6.;t% 5١ 0h`\JQk9+Q,y%PP[>&&M}Lb銄{q57-ă j=xEM2} *hǸ0`ho^Y@&PE;קx wvNfmzmwHtqaè]i:ouOpOKBVEz+Tz?;AԖ$"&q?_j~_?mev_]|27vq6sNh~/>^x4  fQ^/g!G"`Lj`>9(՞g,3LϺ|;yy>9o_B|s03zl^"\tzh5P.D$QV d 3ˍ:nmDKWdtOBL/31(QC{!b4Q*ł xŀ1h1HXdk/U[q*%B!1q((v*!qSWJq j|]Dјup4|eq}z5AqaP~Zk P;y.dT潄X}$B.,s`An$զ$` Iz[0ބ۴rsj>jt>67ɜ׮\ż5ҔXg01|>>qz8}ua 1tzyq{sKkU07ӫĘhGWTw.'>8IƱcUȌ[3zj$oӊl띐=K#':Qk_!Ejoq|f:^͜ ]lfǻ;QPH٫Y$z̍,Ӻ['uKQ ڦhRrMYh Ky/[8\:1p粿jԜEf뛛$o8tKtƘީpgH^6aFDZT[b C-+6>`źiWU1^`/3q̎0SNQW֦); V$!2,O>8SʴqlH99ʞ*R2v1co>E`R2I:a+'Bkk1dvN 2H +bƱJi.+%Gfdc"aB ,XY139)EVa""v{6(y?6KǻTJJ;81y7˯RV* kB 7[m"6EɅBow>QEVR쐔TKB!‹n(hV|ÌGno7( 6K<ތ Pvp݅${VOس +.1\ۅ#Stocla8 9i%v(݈3:\ژW.8\1UuT6FSLaU{K3Jo< 10`0Q&x&#uP'KhSA-%rWx1QLP:-DkzVQVTidq}9_m]Cōir|S2?K@Mҽ >WT&MÛe۷?a%H铱,x7AŬFvE`_]{}O1+Di90yIC)Y0N]EOWm_qnGҷ|7-R`~39*[ju$Li^eu*tؤ :5.v4+WϢ-JpjHk+HmQȕsjKSV}>OϠsCIL2}jt}ϧ$MPZW>}qZ{z)L5{D>>ήD]=wJ6c<lo5[՞u DԐ> ˞a ? ]t1@]{+3" z+8uiC^q9fJ^A2g:g 'zЫ BBOG;nRq7}%mCT*FPnFƉ|fRRp;zHU&Rz‘RN-^kTK FrhQ[G V' ]Tb)nozquŊPR ]BE" ؉h"9*ͽ,R#҉{P" %iU:ʣr֩ybE PFB*d1ph 9QɐqpJăB%Cw\mwh})Ć4ݐ"R\҉ i"W_bB|Ti-/~kZTTL}3B+˵F3F1'P&#y+YHbiH(`o7GM6njJ<֥a|;W 2" ݬi>^2~W0q?oaw?cغ6fRbA3V~&G'߀N${$i&2̍2&*Ϭ'. ӆ*gYDlw+^ F ;}"8ÚЀ`K=cXق#X!%FZ0-5DZ_3yΰUH0W|+gIdn9/PncɐT >T" u]s/tvsLĹIT@A|$aN"ڊvf5n QlJ@\ٌ gJ1YJ&zƕt*71U;7mJH`xeϲ [eX* }]EPa &$[Q׳uPFSU ۥ$Q]"xӃKvrXmH9΢)'(ƝjJ6&PwxVa4u {X /Gܟ&T9ޅE͊a/ dzB #~hu-x$7uUy\<y[f5x?h`F_Y݊5尗48-o5m Y% S|raBTSEXň!UԨ.jBٍN `blpŨ7RYM b>2a c;/Wч$1ZQ6zBښ9-TIMcZ-yN" bcćٗ_7`dX %~Eq3mƻ.,ӭfJ8)]NGȚ)zJMϻ*(;j t'3]$2.7SX"lM,zȚ]GP0іѨLGR6ۧ-IRӁ %]#b73 PNczE@m#F*ʸ]{?5uUk+ʓvԐ I7m(4F'yYHyU&Of~Vގ:D1;Bɵqx!I*k_ˑOqV9Qi:p{ /7ӭ? u[[x)&6uZb!ZTzPf%U{:9Tvg@JsYԁ*0 Zid+9.}>YEH87=o>v+gW@^` ՓGkϻFN${rt9Ξ7v0fEYG񞗨VԳ7 Mұ8I i̱eW&d͑ݺ25]Ⱦ&gmϯ!8A'Sc+ tG3FU0G{hJY,1U^<]_]#63*mŔ^8ġ%^CTFM:шdcSNx,t RgɩAAp6ZӘ 䯅y!b! L>d9 (>/Ki2~ަ^QGmV f_ ov4)foD~98(pZcF I&Ox$촜MChZ7#ybR°_ ^FС$` Md1#TRAXW>J]YQvZQ_a"6grT5MY؀|1HL0"DoKq'VN&BYN+LҎTb!uIk$ms#e#R7($Zb=Ijɩ>rhE.pZ@pz&_\z?.!b=w%\&6u*:XȒh3D.IK]A-4eHCݾ{Xnh3MDKiiAl[vOsnҶW}o.~@b P{SF6VDe3pF5:l`DAN+k?I~deXZ<+L(.Wu)y\*D2B%ŲgyLc!'[k+;JQ$[2"%e01F+\BJ6K{Bs5)l.7lE1U'*%T2iܶj,-T&ED.{m=KIjR6 všwrBR'8Es.#u3^IYظmIA Z X-՝-lpʍ/&[3YJ"v[0 ݳ鉽bjB yy&W[FIݞN]`evAt֐kvNjIiym\ /F:a~96`"9 mMohT^o-LpDXQL[PO >ȣF8!NJBl"9 QQUɁLK}7{9A q5:ЕsNƭ=akއ2gHt @c9q-P@y̋Q7{P^v&ƷW| yL ,Tb_ˋG; +GX8'm8$//Ͼ@~([m No.~7ާْogY:߽z|]oXr]6&(7wo/'}޺B8v=w@ׯߺ>:|J,ԓ(u9߫]o<73&pٷ| ~ SLMQBUW,u7ޑe(zrdJvZzKs\)`h2_#B k2At yrZk.P"˻-+(^.j|rB5[I/ꃉ6?_ݼi 17~bZ:r~BǪgȭ;!HD뛿_C XK3Ma`le2k_OڳW}>-dq?{,h!L~گ]d|LRTP1[u~Z\'`|b>nUSĶ%]5[Y.4hK5:A1[hIW0O,rflL'\Ooܑfu*;яgpF)W*_Z'u[E딺-enTu))`UXW:eqZl'u姵idd;2xқFzQ1a ɋ{l xfwu\^}I/|[rw_5 w_DOܸJv;}?{saj= G;gIA@"5y#b̄^2aC^~2TyuCa=z١ǤtFA;$j-I`5QTK8.kjJ﷯OVv%v dm3E M &]Xn-d^HuJ%D+r߀?Hb%Fis }[h,=}Dy٢ 8V*X-V;+s"P9H`1}&f=QG[SsA8v:޾)'2+EjU6$'WHLdeSbܻ=Eմ*׹*T g7&ˣ'?t.6,-nXp7Wf9Hr ~<{~O_/r xAg,(/ (77#ߝLcS^ǒ/Ryu0צ>= ie.5f{a= o݊* Ynlj%Nj,$_Gn}ePc:}%/.:]g>tthmtBpeS+=QqnFUH#F21ݎYEXz֛6|&MF6t.ICV(^j?e<;h?mЕʃ:O}ŵ1,dǮ1 4C):Ijʱ#SfxGMMHU4IQtNt^f>I:`^I7=e,ss+KCX-Ӈ5}䦧(!d]<֑Z̮zFL?>rUƔ7S:xz-WUrubvdFkrլLA5n:Epb}6v)gȭBuӬ ˎ7 N[T ?>jv@cŠu `ŪvK?@\|uJ 7"ޫDuzt-1~EDZ飀tN,GM&'X <7 wvGDڳbhЕt;rs e)U,fvvb= 'EQ2/.~br Ib>̣u|]ǘCQD!է;IeJ6>s$ j:gPxIH{b~TZc ɺ&)Rw:J"`j)B2 E-fs+9٣ Ȏ Cj&V2,c%1@DQiJ&BT2p3e7F8!E+Prr栲\#sH2!' UGɹh|$9v5yo3*T+ɷۏj9^QbR˜ 91F2J\r4ɳl2ixgwH,oLl$l#I(Ph xgQ7*Q cAkIُ:x35dv>ZUaK9Ӊy;1kПyNTGg{A\Ss@SH䣐 p)G4c4hJlTz}UD\d i)M!@ r>-/X6\{57 WB"^*TO{__Z7w+pz(afvslg/~L F-Jkab\$5,0uҫ)M0qj iKoؖlqj,Z y/vQi}R[9qH4)gZ6҇˭H:_]vvSٸ8R Ą"ir_HQK8h\%tn(JY:ߵF'\J2GF467vTC/hL(P7w߇WE~UWXKxu%+"pR.djQh#U7?P۷o5e}㕒?oN(d$<d;k3g6B]Hffx7w[(T0Q8pM8.޻ gܧެ=$Hw{ںvV*׷rEUmubiBնգ@![s#hqڔsF)k c$F-y<,,~: -djK"6P%HѶ8L>Qc\>%AmK5~ |=r芟ˮ]v/#i[ڋE0[0Z}~l`fØ\\];рԒTqWE QכJ@d|8P|`dRl6SX%4 o0kU.;"HW5.Es!6c-CDYًR&K8y9‰*%G&#H87:$A 6" LD>rY#G&Ǫ"$i>ΣKiuD.k^ݛ>.ATE$58)K8[R"'-FZĚ0mթTX &ԃI)eU+YORgHzGE/\D}d#x{WY>ƃ%!BpMH&Fyzj; נAS2޲AAF6o!$ e1EyzM R'@Fի0 Xun,YēCF+ f77k knw.HLky[hf6܌fQ '8$0A&sv8Tj&a$<kAc>5|7 N#!(ٸ<8 "aXFS"xAko^OJyR3 ېHDCBRr}"1v"vBU,8:0_A5%!)M I%^}zܨIX㥐l£5^&(\˻Byє 65'A 5WH4n2 FAP}׭l*ӏy(g;pz)-)$Psқf|GMg75FȎL8rm jE5!ve޽σ&Bf!%g2NN/[Ɋ @th*/@ <Vf6Ƒ;}R.֏,Y!TӒ;3dfqzӽgUczLJ^ox}b?U%+{;L:竽~{]oɟSEV( -()x}re $u#P  JjH(ouo5eMK%8&4f8o=2HT>~UlZ%GL[7 EPsw6NǗ}=_e?Ŗ.s4J1ȍz&8N.% 1㲎mJFLt}0m-eIW,2A"W3*Ul7>,9؊vճkR:w3b e٫}~k7/ZhvE\Q:suLʅ!rh~O'ٿqZ^IvO3壯K EZ>ƒx.x_S?.TǩmJ4"aijoǁj&NqAL&(F)°(QrBL-bB)VZRYc+~uvmG Tp q/.T#1)#ҐCkJ(шLDeR`Ȁ05 C>2ə[: Ex FҺ:O| C@/dLXȦ!(%DȓKocxU.|S2Ee|7V Rɵ; ??^֋-Sl}eNr(2pL[nD#>~M^zYvbɏn "vy:E_ݏ// jKoCe^\ 347E3^w ܵ6ة)ܫշ;w/JꁽZq.d-%"< FU TG$D{HiW.8$ C)|ظ}O8j/`!x68yp‹AI}s@2hdi9wԭht0;P^ZZݻh0^+hʟwKr.=<ם'{;48+cٿR܀Ox`G~UGLne4AI4OR []Ao14ٿ$Yi"f&_"LS71J"c!UE&&&2@QFߛp[lϘٻ.ځ}=P@ҙ&d Y$;%rAʩ=:U@viT."4HGQ %u+x!FI'm 1L?[cM ˽R*¤:Q”>Qt{"6^Uދٖ 2ʼn9dO)Cb'9ܐR$L!r1Jdֹq0Cluǚ$"f;"#"oI-3oP\>tB`ˠLB~ s B+Crw@ >iw8+A{c}N `UM<9;d8h12gsTωta0B4tkֳhDbsyb 0ve+\Y+ps.< Ϲw RWEΡlTIgcCϳ9@N%~,Zu&f:s[βAfmyC׮ޭvzX9͝km+яʯ/BbVd8vHݶÿ-~j<$]#ђ}<qb_ /s#k$߿}Oe\#OQZO `S7jǝx%T )D&RSM(Hc-( 0KYJ0npei̻AiWD0A7S0 %mJ1k,)0,p0_4I'8 #^Q`D֓S/F&`5da?_; ;xҡivLyF6xgj4!tHRs_oԍlRh En?%>\$M?vL:2Z;ZDKk}\OGn(wvn3(i ٷTKyyvG,g&>_~;$M/vj!('z]zKIt!qoGeW 0yi_6qo">m'~ifCV}/;ƌ4άa;K\rSj@ yj]!iHU*$Z-BI5d;%fNz\Ellnja+)%: m"6u"Z%^Js:.`7`H\\+ͼ )h{0SbsTWl XK-[ gVe+Qv1;݅߰^Dv+bb,U2.r vmf/uܳ@!ָњݻˆ%0{|;:,ZЭ,Zzo:l<:^f0Zz 6H~2{ *~#&857H&ӱZ٤6ݽnF^_S^ވj(~h/vU$vqv/ hdG[QAl]Q~' reAm:Fh~Xߓ:=/)]^ .SFdVwV@ 'ϺH9ݰ&1G)d\!"AMdbE\) 3Kl 4֊*R[2̓R˵ȫWo׉cIF>` ;_fz6njySw_DFsyW3J@b. dank.,՝ڶZ!<2 %7a@C0#(ӰD%FRHM@=" Y)*ﳽݥsv# J(?0KGa1>0"یg*KÕCGYq_ş\!K!|,*[pu㶿5f}Bۑaʜ(sw \|1?ʔ竎\VqX(KB< 8zUiufk=֬e9!߸v)Y7"uA&|.TUR¾[gFZ>4WN:%c=F Cn]1:]ĺ/δ2enkАo\E;׃][7[W Nw4nGY*}2֙Ѻ֭ UN8uA&ykJuҷuh]և|*I8ZȞucD!XQhbι(:]dڙѺ֭ UJ|H]e RNbb|(s ;7~? Q?m%lymiPdVkg+w6`6ӌ]i 4vv?ɇi6>rXC9Mx٣;Ĥϣċ{9z3nI`_L'(4вBej}^[K ]k I/}6cE&7!?k}D QJ\#<{רY_>0C[<dq%og=0e] G_4{|H寯z3o+Af9qkU?F>aOޅL/J,?2-,`~k oF8n`.FXks:[W0{m]p:Ya6VM$lϸ XL]4>XXl%,`q/FK"/sDv%5%VCmB{[~I,6*~+&OfΘ){~#w>uq)&@j PO+:@ SPc@a TrѲhb, ťW#;f jfP-Gf|.m JZxըN CF@^}2y=R"LvLZL:0/``n`jI9!0 0Dakb'uP/XX RԜFńJ*D}ё!\!rYH=/4"H\+a!:C 9{N1HU+-Bwخ#1"FMXc',1Pwc~jskٴ,|:K˕)Sow bJmQt7E}$p`P_,)bCŸb,`nkY4Êߋ1[BFgE:g0^ f/Q `8 Y`x@/)a6MXpD8QRG"X1E<\k6$ae|&M'$rR0lJaޮ/U lZV'/>ƻӨ[kThK3示Sm% P Sw"* 쩣F"R(CS#5m> lUa{`.؉JOE |]7TʲK Kl2d$.'U4/.G/ZQ\....#%kNr"5A!J ® !"-!$&1Z1#8B" E Mp$ Eޖ22LF0IC빛$l"NFƈ}%F( P6S eP`=I %F A Cfm|0Ҡ #nQp ,X$q[l6cwnǮױ迧M\:Jr1_.vp1x) F돱->gM4i pcd{s̮:zo ZqXi|JTV`h|0&_".Nb.h\<$lQTGN/rQ!/F |xN"+kΦc1xp @ƣW_tVapDl}b#5Xc!h@4"Q2`@DAB3+s܅}{⼩]ښ^ETzg\v@SڲUg;PyvE:Ti8?X5"RA(<Ѐ1D:f:gu( _LME1MD!]q/=!*:,I@ްnIقyߕ^m`i~U^#•_iĎ yGdJO"ryj- _95G KUX )4Փ_˶& !,9eRJ_ rL!D(:HU$uBe`>M C|SX7_&^K}L}͝gLFLJjiR&+N*Zî f^Qeqhy<.Fܦ|.Ѐj)vS k”H[8­,i_piVݵnH:zu R4d Sid`4%Wץkezv cm`(_\Qq~;*xsSF#*/shl-`R;?9A@SʠyD> T~ Wy: ܦ Ma/C |}9@GD+i*/wwl>zەtr.0X5o+DZ,}5:^̣u;Ǧh1Cε+T\uU!QOpxRiٿ>6 x)/8 Tz0O+FTyS -N<-|jI%wRѬNШg=syX0ksˀ_şl&T#s.g<|v~{NG]rЍ}xƹ AiFO\?bǥCO UV5[zU$g IND܈lՏ|*S5b׀7nđZoz*~4 *;Cj$tB$جvai[_#Kw~:|y1e,qv"t'ahO1ךjԆ⟜sUOũ@ΕϭB"@ ~ʹ zdw0+6Njl滴ljI؄1$vw$8og@qwq&~ˣsη@}Ҕ(N ]ўbPαO#fرsOړ$D;Y /iFh-[q DI@:v}kZl 3f@p%r5Ѳhw?gK&|z`]؞R JZC;Zɶl&\!T ms0֟6Ƴ8IMt?uiܪR {(0![:rŬ vsx.d<8֧[K͸zųi!8C5_DLY[Ǿ/ܞ),DA*'[+g:ړO0~'iv7!6IgkNI "m!h>aƘEQ](2kOZ 7\ASQOXRZ0#SD^ulЯ_xu`H#)حMS~m3%-ݪrҲrҲ"$Mp==U[jF>K5ֲ&r ~=䱮Ѐf[f'Y"+c|fwy8d;tWôo3~WaFC[{fAm6۝yG㊟ P vsǗs3 6ت\kẆSC4Rw׀jM^Wp{ZmVUך'mʉ v{c$[HA x] c:v~ӹlO4Xrv:wTL0c騣by!G%>kExZ8hP04C/.a?-N=JJ?XW&ȓ&q´o-'\3,U&4#ؗv0oےb>,$z ]l36 ,OŖcȑ{; γQ:Q:Vs%;%%eORS,$WWaձM= 1c! _~I<02m֡gn;[Vr:9tJp XE~uJz̢Tk$Dާ/~ VX۝zKc^\4^?ExhlYaՄf)!^Z}2l†-Mf?ڀr Q'Uպgz7_!tDv?7'cr'0(5#Q[Mj!Xܚr|Ɉȯkџ H8ls1e8t;K{? %o+~k.MפEI+'Ϝ`>O'cd5xs o)\#RqmOrYOlWqeljUKTR%*O:MmJihD6-F~{EH6Q'xm" D{a|0$ A, o'/\ڠ1mK h mGxrB1eyՍq( ZX[=ɮ"e~;3Գ/0*+?y 12q {K"̸ȴx3H_-w;x:GX؈6~vʣeY/W~ARB)j) G `)eML-1ƁkJ%\Ff-ntQ2CJ0w͍ǣj 7˿T ŻϨXvhzlŊ[jWRsJQMGo@5}5a+RRrRs[fMdhevydY]o egPV ĮC#mo)VB@l8};+U!pe+Axgc[rM^s6bchn%JpŜ4 d}#l/ &Op yK ( ޅqy ֧c4eYg9i_H.x|6J= U~|T/k+nJUja_j5zOL1*&+UQmX3x>b9r7oj%-߄3b "~)U&mUʻg;0.i̥ADT/FC{G9GvK+5UF%e9 ,[0% KJخB bFpF{wӂͿ^"ƈ~xf4UdGޥ]%FDE+0I7T {}V ,%>8foKk-l؊/d6s<7qrUh\/Y5L!_?X}n'I4=ﰹ˃g"$YqsM%jv@а:Ju(Wk}n(V GRɸ !n k!\æ xVuĭ6](?C )oQء0WYvс` }}՗/[FϠBZ5f5Fm0-?IpL+0o#ݑ9˩g_1 tqw`Iڰfw0[U'HdF(KT@ ρ} rH [,߃3\z8"x f) NbP mL"sj)z(:tK0``;>C+LCuљV~aEۀµ"UTx< H'٢룕{ i&V~ fɖG͠ |Ս(Pu3e j-/d=JhʝƧLfY2n(4o,-CF$( -ښZ36x|%hN1VȟȸܤF 9icA-E0FUDrJʋt J1 n׿9'f8V料ASoD`;tɉ2<9ydf\^ ~=f>htBk?eC2t1Y%ӡC8:}?X oYsWz'`X  Q,@Ԭ;RL\0ek{4+؉ e-3C31}4_wR.ڟv6յCnUШ{^]WRv(\{뻞gTCL{_{J;fSp nYn>u ϼڲ,ug_ɿr˭puxdbIL鐱=| ];p6|V=8[z \un9;^JyK𶦧Dx[zTvnw]}&j'=Ɋ9h+u֝LuDZ}by2FpOZޡ,p :/~N?njGޖRSFt&ݪU!E!h<.tsGɃ:XfCHV'"yD4DϸPlihԟ^! /]].RciC'SC)DC6á^}MGD̺_0TKK85 G (gO 7K4ԡ陾כ(O't'sjyB2x7jm~QF~"9n3T _Vץq$H<pFqQM%3DC TƜ `L=R{`ep >tjp4\n)'Ae/75qK;^AH1bUfW/QWX3.KOco&P-t;3t%SgA RaG7^tŽLBRrs#Nrk{J9.{Gt3س GQ,VA`ڱ/0"C)̊n޸Q 1V?ri 5nuJ1J%qWL@ank YaJ,t߭bcy^8&Yz،("\ 4q`k0]Ŧ|lk:isgQ@UI S)-PgD`d̟SO\ f} ,qbOx<}/~vԯTp2TU>VZڈ *5³̹B'ΰ"9%@N, 7O \X2nヂzw,s[-E[Sǔ[3ֲ|O33뫢%՗l˭KRAs!Oy Pw1ctfRs * VKb0`vDyĠd]ٰW[5eDu|z*W2)mElŊV֍?Wd| qTMXS߂V~@1T-Cg3,;mtpbXwl8UmĖfz_+,$*6[2e]vqI l}VgMq ;2*6I܂WkӥRogFH⮝*uAD?j.XX}([N837M4|,:ssxn;}st5"LeiOK7"%"gHW1;K~'ajeyC?8|ON5jF580Ud`8'ȡ<ԁPZZ/~kĵ}i}G`3Nant=} w׏jA21rwNj,:sw_y܎0/NGGw<;qbϡ˟qgF sxP^{;^nMU'MbX]1AB< ^r-<.}j8 7ڵUnL|Ga4iǒc޺3#nQȽG)+QMry\߳ u bPRdŢ$ %Pc~L߂U*p$4A zueHf B[<|y2#Egq ]$(E3I\ӥ4p v؃t>aJ s S1ԻجQu6AK/4}GtaO89N^RJ\;d{gxZ|^OLX}v+?cd, 5uzd ldSɬ ɱ25oi?`7W;>$&cƥVo ),#%RO2l.@ ] +U.[ q`8pϜi4{Hm4ōNj/WGk3xg kn22.HINy⎝rԐk@bOџ{Ǫ<͟7K7.W {>>]0s!tlffY)k_Fzۘ'礜.t(d_ n}A{` uC,Tx|枤PS>ΰ)pC=wɴѺfi)'s\ho63nbZ>0? h8Rb#OHoTnLIre'wɽrK4zsrkfJdئ9%F"B?U+m_"-p!M&pQ.[ fkg  V&\$f{*$ajXwmmrUzZ+yI*.\mںEMIÛ!f#j#uhtshT,RU)cBDa AV[x3ڵjqF9m&Kh_9^$vV쯻ӎL SRu\3Hh4'izvR%p1YDh@`N.rhuCMdh`FBai*H찠)ϱ1X΢)UCMTU-VkU@{U}q7Kj f׼r[Iz7zb` , q F^>LpI#s2nwưx`928%LETp& F(EX>q/5D01f`WnuAf" ~J95NYr G$t1);^:9 aVf֢_yʼİHE`8Og<(Bo 9npOFHB\Vp.n ";a${ L͟/9ڽ5]V֩Nk1(t nn=`ڬj7㩨GTvE6!Vr$&3`/Ӻ D)v)T\xW[$w+IJ&:eS%&|6VᣜUWE\u'=5Q \K uCxv8zJƒC2xGbu~ l?-;[+R"aMKgԓ/SάbԓJ{ Z} Ò %^L Z[sYhI7yt{B7)fy[z~inHJ=;:*4Y*mw4j'p˯ });{dGeHmʫi)jt_7jk;Y͞eՅ\ܵlU.J0`2NeDdz^ S`MfvdZwoVmRHu4'_hwA/AYhE{sr I\ݫMRQ bMn8Xs4%嵄 iD UhUw@'j)hY7Vt0%m˶\Hu-x{DI.'2.N9_J#_lQ96;'oM3`u|`7Q!ZC ~JןOFq|{7;0WN[\MFԟ0w"\A|u7-Qhw'N#v2f&wQNV"+<3sp|*õ84Sޟ2Q:IH} ӫބ˴eEPV+0kn)_~7[/%ξ aʧWSïէ`.&R"j?6-ގ`\.].:%?+Ynl/x7CnI<FȞ"zQĈMe]BdW\B<pw14ʜA:-h5Oh A0%1n-iտ/l. q.G~Wˢ2yeٽdxUVsߛGPa݌w$>El,BYSE`C* Sm]EHY ܕ2P#0`PC0pMLD&(!bX\T=B&n:3 `j<gǷ"6dلaP)`*%PJuvO6փҀJ ͸0_d3?pyuS`]lն٥lɥc8gsL^iTPXb '6jݑ YE]5 㚭Sx&\ 9+D4"qA(F6]d\&o$ZdЩހ2p:eVRiI$e2B_͓"Ydu ΣX*B{W {*p"ZȾ A X5eiT!VX9<1f8-IhAX| 6t|=R" $=^680ܳuc9i/*GzK+˘&{IT*TNzdEI‡Dv5L1[xD0rH+rS#~7أ xk~P vE=,lfBVr:QO4RȎW{Q=;qĩ.e>Ol>9{5ϊϻ", ܗq/oFqTz>zYՋ+['32$T۳!ܡjI`2"jtODZ]܉;-N W'G">S"j&6Ȝ%Rf5_D DnzRk9׎YGJ]S)y$s'ONS%OX=l31'"4IlRo.w]OI;GW[xq缣kR) O|9wY$ʂ!C9 Zg& eI}0+Iq%$ϭe#-"vRY\tTy*وZƤ 삟ֽSq0Z)ٌzFCJYkW9-}5 hy;_rB.,گY!ranځI8Ӌ[3Dɮ}X[[Ͻy9;^ۼ[V,DYnļ39э0Rso[ΌUAh T@u~w~Yޗa'.ͦut0@Q֠Oe6K̼Sse5ʤsf{a`h@9 ] DNJ[h+Wп0ItQH@? +p3? 4?R ;;յ)8NA=L˥o`gӂ#utȡ..tzX0- -RCHZM4JNٺk% C}{Lӵh%Ho2|> $C˂yCT B sx'*Z'KYީy@h;^Fg"`"߻RMU{5-WUISU+08 -Pt +Q9qDbc17k'\\U!)@ 78e$P&J @w!SG)+CU|Mw R[>6|=zO$,:ѐeyN?4*8fIDyiz"cO+$l)H)9vt^4GaHȧhQO/=R^'/6 @k7ILDTrK,ٛq]s"z1ñO3{,&F)rm콥 C"^De3"B 9|J=UAR=h(lYIÒ1LQ JXy.ZWÇ`I2hZQ9mkYڨŖH]K/>], H/m<4˛L!JXTy^ìP-wM%BKǑ}i$TI/ryNyжv\]*#X)黏aR#,%YvhMaPCnNiMᶸ&݊'+EKx j!K7QOq@VADnדhZ'صt+htkC~p-)nIt[x#p׼} Jn8[S `EN=uN[2tk#`',AHᣌ\k3Hq̄xpeV ɓ`A$jG׀ kuun=o(DDD'HXnk 9X1`4D"):h˜L(A +R YјQbcQds@Sh*doMOh OQ<|#c]JR rX&퀆#дvڵt+htkC~p-EZMAnNiM!wXrt-݊'+Ex)݊f"$s?BFh=vc*-4Fا R&`6ɲ QtHx])d+D#^$ %\H !5$ M FI=3P9|m|(dE2(ph*Evŷ{sZEKxJ+z$  |dm h8u43[e&g|I&i>e|W5JakƓ۳ܢ_vljX_-VMҌ=.Pp;?qvfCo#6%ޕAWՉ_v3w޾J3 GULp$ٲ}#Ms袑̤!XȀF^b?t{L~$f V6*3[oA4RULq2+HRKi$[i޲zH2+HRI IAXAھġQUWڢ44>Ր򐪫G! FR}5DZCACA ; KJipÒѐ #I/X(9x4ߨj 5Z lUR}<0&lhb'H9Z{64 äV 6۲D>}L){KpRbfR<;ˀAkjWap.*ԯVu}34yߕ{m+b'U W0vy~I!~~ٷ}>B]XoL>hCaw=M/a<(Tt gb8wHzWaV DzPLgҕʂB;k]nBꦌGl[,Ȳsˢ$hY,H$j~IT(sa2)~*yP㫨 O#Ԛ9\<\8m0eE#2hY,F=Z 2` ˘X1aBHNd6RZjJO`3RpDW3 =g{ǫMΤoF)"uDuF 0ӏם $Wih:ڋ4kkQU/eKzѷ!h\VN̔0҆{DesAk$Ja3.Tc v>KVHb5s<cc*M"*B[r5sj+zUaFֲA"X)Z&~W۰M>3'$ctRӿXD[ﳿ0ww%@t`6:bɥ7ScU!Bћ ۗ+aXA/oEEEE_n<*JK| Yl!$8YLThdITYrEEh>T$xzy+/Uu$ur'zә(c)<`浜agO37+d-W)?z(f0u׀Us\OFaJ kK#cz]jK,1cL rƒRӷW>f@5_UyZ s$DcUwԗAN/KoKUJ;oogcj4qaј'hԈ1e~A)#G VF:3%6K'KJYA׈4 {⤵CNJ&ifi$a aJ/eFk%ZhgV9ie)*(({$6!{H?' qZ)m!IwيuAf@ $E{ 3 cş‘)`ą*ԀD^0WS&7/@zEcq2DVښ_ I8n|JwS5XHavl/N5pFh_QಭW eTdl_Lhv4I /60DjG{:6H+A}iZI1%}ڈ!-#W-H  iB:X)6:%BzW/ EVƑ :ho1 Nq z+1tTI!4e {2؄UR(_ )FD%qXe >'H"R%F*L~S_x xas>}G؏_kQGK}$Wƞk, 6v20-"W9r'ʺʑ; ;>܂5|{XяxXяaEK X$Ԟ y(Uf@Qdh3ň!Ashl23N̹̚ 3Iaw(() {m:Xq`\Y9+K@iUjP(üiR^}f!6_ ؐ"iuAr(pHɉ 9s0yZ9fʡ肳Lj9"eI̩[CQƄ@+e&f gs+} Z򐑒PFf1," d1.dadN7")< XlǞaKJ QܕR2bD5Ti.C ,ָ!YT2b򐍒IQC "pٵ 6rp(t Fնu?w~lܧGQ6U?_mMΟ~>|ˇ~@0%Qw#n jP}2/OѢpo~:~QĜ0oCW}'LH;cJx|T LL+GS=1btu;Wst3PE' P4Q T,1}9nW %9ef*=(%8f\̺86j8)oMŚ9n3Rƨ?=M' нIIØ~8# 5?'MHvAuy0dzm``\v(-\S6!F7{Ct|37 g6"jT/uW>j֩bz[56?j$k7$;3"ct<=3 ѫ&8fdk*F8{, Lrx|qyts{NjrژL09՗Ǜi&p܌J5;mbx7l:'0w~xs>>Xt/fxOf8nӭq2AQő7EkIuR#왦*kcqa(wfȈr;i- W?:)W/U#h}jL@6޿FfBη'f2AA$+t{"qZxu®Xc$CFԂm90|v(~M!l@X3uРky㻢#=4Fǁvb3zO0/U7'r8q3 ̥?f~oLծȝcЛ8 -' ">t=ۘd,g̅aV,rJvcwu:ôFɌ΢ <y Jȭhr!)-rf\>pZmqENpE.kL1zE.K3eKXNRM"WQ! ^HIzT &|+rs֨pjLB1ARd0psRUsNns1-7:~XxpcO/dzp$K-3wR0h(@Qv!͖wi d>IE|2{wv/N0 nd<̔"x! %.(̝Y7N!TѠDR-ƒ@2 l+R ֎V EXՆ.H=FhJt!8+ = N5|3+Ǥ2~_^Ք}~ߜ t;xUx |kvw,!xM(esY_"k0ԩly|Pp# oY|,m" 5 ) B !=$!- sAX{ 1 lZSa1N\YCJ4MO&bV> %(VX*P>M;@lAyT\<̉Mɝv?rnR )*'~<_Oar3do?s o:QpӻCjt0wm$_CWU M$!d ӖWdI}O5IICz8CJrږDTU=5YxL4cim4;С7S×zmWՕ<}8/u٧|~w 3e>2{nyO7Wu!|ni'Gq%W|݁< wgݿۣ}$'ĎB׀:Gh'w/gLu.T>3TwYp~vGwl}mkz -|՝wXc)cH@m|bj[3j/hz~-v $cBwSc_+oC˲z\IoI\T4ы( ֲM@X 2M@)ڔI3~I=,J|E +ZYBӌ34r^8l4s>b`4FʍR`j jqq30 0ee[ Tx4QvDRFL(owG.>&ö>:pA$%z:zXcM#8fV$32DXNɣO$Y]ݘ3J @D0=ǮOhb-%\?Z)՝h=2ƕ'7;f; !HnNE*tZW?W.יi>B$ :93# 3~OS.VBm52L`%.p] l AŤ7_Pi +!"%o =Bi5eSRs3p4UMh-XjxRUb@*D~dy[Vn(uPpnQό@m%KA;XV6?B:.+;YŧY6Ѹ5}5+ cܠhAːmy΅H8[.@5+7zucg{})mKAO74`'rm% g豏96+gEtWҝLgAugv2DV [rnJqTH1{Wy[O-oP)6S)&l-و,AU$܎TۘcITT;\*de$YcY{D^F),T3툵r..%N:F#=8B٦DYc6H6 b8Hm8Hp +6d_AlTʠ4gA o 'F H7A6Ի៕⢼Oľ3OO Śklh'X6,5i<yC!J~s㟦T4Q+&?) LTs!)tƬr"Fyed̹(%t c?cGƺZ5??VCX 3eꆍe77w%MW g~v4 V+y %P?ڟ|s4^rDʫ.; lxs;mtY-xzwɯڜR;ZiMG[GbHg?o{8AAcي}5-Qoll?7W[\[t{C'h󊈺9+/onsݻʟ?O+sx}vemԘ&Rc Wb@W7 :P Q\C $G?8߅OvWcwc_tr ^ vQxe?-̾gi(i*z5"Zb cQ\0>G9҂#:I%B8N4ɋ@Q}0 8?z5}^ۊ+rSLP w;r0VoEXHGvG\+6ܪՆ>0u(QˍA$>~u(TCȽg9i@Kn]F֫Lʨ1 94mI8 nҥZ"L7gF/bs})(#"94zucG%Ka+z[c:dwͣObSwHtQԑH1;ZsYfN[;v`LJ']+Y&aD_LkI2 5cYhiw" @8#Z3m"/<%Џ:LuX)k1N[-sP-{AѵϱS@v:/D ۛ2On-BCQMLWPª a*oڷ||DoSI0ړ"CR~\_AVkkՁ#G׹8 3o] :w?0C_1GWG?Ӏ/CliMҡtNo:r֝56 n;Bԍ nyk谦$oLw0W1NB^dRAIgmy\}kyWӾMyU= || }.wJit]"NmyܦqzΦޔ{._};dcY7u5pkjkބaW B&pob`Ж%orӭ%RIZ-c;UˎF2 /UVrZi_;KPPKuOΒd& L V87`ЂlkQCc_>?%u)ӪN;J$*!B* h*1Pf4$ Ģ$l揱/c_ D^rNJΝTjUgK 9UA6RXg*"eRy\0rJ:&eJ1woW ;bi|VlfU$rnUHM8T?BR7"gH Q{.Ť0*Ccؙ!?i# P} [f̙g8$Ɂee-v>[v90~j|C^cҎ"`8,ITX3ÆҔhCD%<:wrCQ'Op30T/YOw52H~{&/19Ov8RVb NFSHa ΎhW ‰X4k5eW%e ce"S&ę 9z8/dQؽc*_uXY@A2{[L.wyfJ(>w3u[ g5aEOrT \`Įʢw[+6"마rEvR dB"9Ap1:=J}i ǏDQf\I(v-LIb~qd|C4 ݎSWg0Ci>;!q:_<\!uln%PVlnU57L[NXj]U(/5ӛ; [ncܕCx7PGz0gպ#|{m&zuACwpǿNcax*F蘒mNǵmpjՑ0N_usa1aZmޯ#w8_f,鵯1} SJ_k]_;D_jyoj%fe7XVWC0mnyˮԼ_Nixknf㻅^4+6e1ix/Lni5NCN=p1}\NDk ʰp.CN0nVcҾ~qeXnFs~Q3w\ٜW:\8P3ϯ4TuqSѷxy+0O͇؆D.2W/y2/wJQW8O{( ֵ;p#SVh; a4\qlj|u6=zqֵ P'ݏBgHfv˜lx4a^eQ'Ϸ{Y' ыnsGCGu } h/JXfW5z? jb̜{g]!-WO7APoZڲmP-l6 z%]P3yAю(yC,_@ L7!|IQ8xM6+!FuCd+ڼV( ҥTQP` իէcBdmSoTV0m%m @)v"!.ǻw>?)N:zQY(ko v6Լ_$0slk'ުP֕7UF~/#?A2?L>ɢ N?vqg!x8u{(T= Iǡ^HJTmUaGz-sv Xtbu/F iCC 9[ ⅉ.رC]З$^ԈP*s(SZQ5Ӂ)Ygt*ز^T**zߘ.!+E&V'Z!ȫ ^Vxb1 yKnYc*\ 6SvzeT"bW!f^iB"@bڂApс٥=UM! UOtVG^LL45Ц9#{A?:Oͨnwܗny:ű=x$5s)4VZCb8aj"R(UVDS7?4OcsPɸ{Д3R8\78[90v'ea7ojO/d8nk^#o ROڃb}Ɖ뛢X_`f92Z*)A8A:ڦ47LX*3[a،ҌS0,~(QGD˕wykRsvBǏX#5BzH.Ƃ'R#ö's?|T[*ʁ6e]il9X赗Mb@^rp{x_V.j}IH֪vز-=dc5GdKܥ9X(ֲu~iMۭGNn{Ʃ! Sڂ)\YQ͕iʭcss&'c8Z/2zRjɍ1dDŪ&j'Hʎa6V6JEC$ĝrxCώ&l>a5D۱'7/I:weCǭc„.2{~.~Y%Q/)n@WN%Z)5oj"B٨z= ƚ0/qܐvp] )MM| ś=e'M@w|^646ecg$v r^7qo/aNyS1Ae2M1AU6x۝Q%H>b/7ebѬI-!ʐ0m`IM~k|crrxC)[?3T&-I4EQp&&R*GDn4RPHCh3,;]E@=,Ľ<fB*= Àoop2c?Nݏh8o7":5F,eIHrbJCixQp“49J2Ǎ~f`_g3ʥ(;mr~Y8amq\f.U?Bb~`y7~t+${N?Ep;p]2;0['_ /|xXxw)"rI"F36Jq*F&M:gKY7=YAwƅk fi+(?p7Q^WRL %WIb*M2raJ2H1)6DbKVI$A˄psx@ t)[[ۻ&Hp߿^CJO:ލ]dXQS ʷ7)OAn0g$!LV4Or cNLdIFr1#k_=zi|-g;n?fL=rq" Q[Q{|G S3H zT2TY뾂3bf2c JrD^N 4Z đe89MlaAhP(rDJi&)(sfD."669࢐sM()2k :6 5Z@$3\)_AidL"*sS.,k3g\uZ.P Ⲇ#LupD(,&$%Arr-Vbꂀ ',\X (` F"cXFRBJ#aifBARz$DcLMii4-%@L2 9;ϩ!Im f:ELr h)X Rj(I4K `6[ `F}@`(A$Ö!ICG$4 VP#^F)SJK]]+ܱu $z|g+V`i +h}F! xh^R%w DvÈ?oB}@ٹux9V)W.'3vpΖۀҦ# f6MaKCz;C\fVd0Ws9PQR&ptM h@r̀oDSK$e,FeSo>b¤~'kgcc҉qWٝYs73;t:fz &uZzcr;EV7f]zmzt^M|s<_tӭkn>,/Nl?vK:>EJ,L'f '3ZwoەtX''$ѳJbڿW EɌ_ H:ƨ/e$ѷ*S*[^uPe;"Nfzw[ҩDJ^[]>z<41yg}XQ?ru&ft7e1./CleNܠxL8 P0%ȧ_. 0FIyuٷxuXpgG+A./se 6A%Q9ʿ[zH;lJzuHșLm1hcǸ 2F.8Y7t%[r"#S;]ݶ{͙:X^ll!WΚsT(yRg\p\XҞK}HKWzɔ8K_ɺ T8кbPFt|ۨbZ]N[\_o.iꐐ3QHunH5F%uĠxQɺ$2$ڳn Z$EK2(@&MOgɶuLG$&}֤=?I20_+h\#jBoyR5ADyx^'Ze݈zbPn+ٯjюpG.mqOeY:/!g.^2Qrtq?يgjylZ .xNMBu|vdz ,H BwgA\Tςij ujgAVdz ,H$YAp;,9qE(I2 ⹁Z5U^oOspndޝDxn U(:Ql1|1fA٩ĉgc1:4Arڻ`Lc1 2_fؘo ƘvI1f&e1sN+2 ms1ת d(dS.c9Ƙk*eb\c1\&տTks_Y`%b9ƘANW21c̵jG1f!11cU$1 bc>Чg c͊1|1f5_YRb9Ƙkv:!dYOYĘ%׼1f1cU=f杵c>BcVX1\&dտBs1ת TŘou5Ƙ Ƭxse %1cO };d6w?vv,>^ /|x8?-{1z]|o`ĉ"rI"F36Jq*F&M7co_7fq)𳮸>,Ɩ0 ᄍ;qϜb ?W1]4\}gTbZkM+-ly,b*5&UL"ksPJʌD kM$)L&4sb S`Po(-\`c,7B`'Ȟe>_7g%0lFieDi,4"-J1Ѱ6NHYq\ Œ"AIB5\X-VrLF%`zٻ7$W ;Z+ixlad+/1,ţEVnӐ>_d BiP,2"էUT U[^B\_B"~CȀXӰ7IacEy/dfl8_twoPȂ0#H@)VJ l[UJ]{zr1+TZDzx6TB1ZN c5&BLtRN2B~rǣe 9=uxRy筰R4p:L(s=KG !:&`\Qacu%l I6J 62UGt'gFs#ERh-Nd:0j1^kb2%5e!/-8 ,󎂑Ydc'zjO("%Vۚҧ`N FQppmS7l00GqT MwUu4V"G aȓT~I`2GR~t2( [JF.O0 eEeG3 x[D+9D$1YgRş3`an3h<{-bj&E֚x5qA%k];8lt30p_c>"҇Qږl] f?D!>yg%`n|EIe/]TMM ,-o): 7;oaN.x#o*ATєI?Z0{*fOES!LҝkgRдՃY+Ė$xZ 23ٯa܍ݔzWuׅx2LeTHǍa(-߳b .s/l`f- 5(BqI!*sn>%k n =|*Nzr]՘6m=kͫ?q,;. P4A?,7/7K*ke|4qZD(qy2{{1 o/x3~@s\D; kKFښ<2 ~!W_[j/_}!o/jY/Ҵ =MIrܤfgn0ܛu8Kq0}$|5N3ƙB6V[Yg|:TTɳ*> K@7ų=!>[f^'f4 {Q6q9Kf8xֳ➦T/ dyxc GHUQ9gwT[пgI0OH,Ұ~I+bWĊI -66^9|$ I0\a7sjT4B H)rV6 nYd%j^B\e/SwQĝjPtWKJ1(Tנ")cVl~CVQJ^IF 8LTe߸ #9ڱ.&6-}IIQ ʓKȯvJOYFSw8dt4>M1f:]]^b";7tUSuyLozmeS7V(/K`fa !gu@ԑ^# olzOK-+u+db]s'cr˙U9pI_T(d&t޷ 8sq֬T}=ve A^b\"'=q0'̫YTQ^O]^ mMM~LKY;S&g{q4}LdGK{/P)}wE Yw>.x1Yo_:1J{z) ">{X">zp /\%vIW,fDҴͮUP֒v@Zk$< t40zSVͻN|.d~ LZ1oQʲRb T4xDyf}?|"ŃzЌ; cT+'q$rO=c,sYUnUHutg#8G-+:f/T"%ThvrIebpޛ=8N㯅cSl\UՁY7|*zQ-|#MMĺ Ov!JjPt[2 Xci9Xp.HJ(2<h@2zqݴYmH>"Crf[y ,TZ% IE=E@AHD k[HTEv[E`p1ޫ9 +&}T 5LZ\)h\(DR%8]T@9QV-Ad$) Q,'Gw F(brβK/H1z%A\WW qjV}M0^DaD5ut.v|0 /rS\8s`qaеj]ރ b)JT얌"c;Mp?4a0͇V`g܅n{%A$+*tq0c|;gf8˨b>: &\tq$VkZ*@r+#t?OGZ"}r/r1B\IܟBih6 'bz `PVś'^FmaN&z5˭*7:$މDeyo\K*3> HG]*eBv isFawr |]NocӉN<{2DE.WoEbiT;zWڎgWUΧD]9ʐ@'_>;{0݋&_>}W ^֭P$u;:Py/ш:+D|vf>g0krQ cB0ڪ=aѩ79`yKs87oF8}'8(W'^!*#:6g}QOZ6g( ohswNCi'>NWH U'P59;P(jn㵣k3n ru.v#EVҫͨKwސj>:y$.eJ {۾0dI?$AW(+e4ԙ)61B "oRөt*:J.JY\bAA +=pA*"`9^-JxωT4zPD!f%j^B\e/aԯt.nҡq,~7-̀ѥASGk҆b|m4S49`W]@yˆ. ?{Wɍ /8GN?>G\Lr!sL hGbqd=\f篋s+zN>u'4MxMV})nޝ12P=iyR>QGcɉЁ;I?rCyt_|/fziFԒ[OsaW3V, OX"ƽtәk7nuf|bEыjz(Q)S> RE6 L_5rhuɽ.SkE P-oUz t(@I7 KBSWJ1e_(:Ӓv:?Y.gg7}ͬPQؼcWrm@!՝!Qz.@}yW穿Vi/W~hٸ="_\cQNFO\_PyYۆ˨r̞p81"16F0e:ȡDhf>l?|I$A.,?Kސ )mOpGMS,$KF!yЂoKQb[#kɕ&߀w?e êf\*2RHRYMA/2d74\/2ZdS. 9oLFRILIqMmi-@.z*p62FwJ!E !KaX,^BKp uֆM 9!k%Y;b-a_sGn $R+VHc`Rh Ƽ<.)D\0c4oGU#*ӘpHFB0 rpݶ+K~1rF^/P0q"s;ݾ|-Gd90]=ot*dd|0t11ҍIUcM]ylj^H5 6K<@+zhޟGb6}wvƅ0gɅ8n:;)7=ۓo̓=@&%}YsEYݤ d`\]te4<'.O% r+˾&锸j"'U:b&yu%)!Dt hְ(|+D0`󢖧N-dD˱ k>Q9)F{\GeC{RH>94(DzHg_[҇%qht>롃^oO<͞;@wl;@!Y 舐쿄-4JPzߪ/-1=B(6R_#·}q>ZqD6G;B;H޻j`v^G` 9{URm$J#VÔ!꺪7g$ԑEё gHU>2hu5`kQ4IoUÙмgEXMpLEt.d`%-@5c!F%KqbK +Y;H߰ua+L15SL-7 z95׵VQK然J􉓢vXs"lX`,n3ׄMyr8&GOjd䨾a;75}B17z?{Òȕ[6ɱ~5W=nBo>15ܼp-ir J'k%lB!u]$&" YC65I _/Drض^<(;r:QPֆ fb5_^ڸp6&l,g{(#OZT_rR00+tՐ# (4uyDY\ëW;RߍՙWai%&ƹKakE}MSڱ>m="TE|ыlpp ptZgU`#NG*EN?Hz@(1Ҝy]*S]{59 ,J g.WIiNO1[F9nݱ(LUQr@6ƪSnpGs*IhЅpz,G9hAo_!Ct;s.?^, &E9(rޞjqJx6xUGqŷGp;=ժn5LmUWď|77PRz sҮm+m:/}`p[{MQ}c 0Wߩ8 4+[sUt Ε@#f̐Gw_ZNP;O$y~yqNMop+kZ'B !j'@2(f[ܒ#]qgIZk.,&| pHS$1:l+1D9)Hɡ#|$v ?:dx-}ຸUc>[-k1qy 4a0GķXb o<'h6^b vU:MoPaѹzB ury'/nSWL[ Z ƤK9)+'e]aڡ*jؼp^D`" ˞-Q_q;89;e˴){17tھscHۨ<,1e%\%lmjM+-F q^h7Aqƴx9[Jr[nݳ-L, C UðE'8Sܥ!9 (f2KH81R Dn0Ӳ{ >i$0Ysm}<܍-|vB3 *3>-oBF5ܼ7XġJSe|`X+@ĞjkmKH@? =`w6d<nȒW )%")$(֩ܫN:+D9K!yosn{6O$C$o3$E_ۖ ~͊F>\Du~켛rP:G$Θo8jS*wΩ2BĈsЀZp{B9GIP﬍;_Rm$bwT[#r1,FX\skd5R an cqK(֨?v#)$I  4xǔ 'by [I=l$ )= /09;pN\Yqs| ~>@Yka1H6!z/Pe1f A?ط`Aֈ kF'Ŧ8|xdE U(Lt=%B4qi&,U ő@@p"MJkj&w):!D!#YļR8) [.(A 43-^=mU .e0՟/H%X Ad!XB8]qJ`s&%NI{yØB)OqpnF1Rd5\^aOeO7[J #30X ՈH*Uم;UҔD\'Kl ;RkVBk[$l tBL 7s> kXo| +fIJ;݀h'>KdpZA9,zbsM# |b9 \(рC3&5l5(C\*!51eeP: hSXrj1he&>ndȱg=Jmр~@yqͰnwp(h]0+Ɗ"ѩ WĆ CUB䛌H"4:FQ* cɎ;*Ե qgl7Ih8AD)q6##5ûEAI!t'd $l##L4Ⱥ`)CS c.z(E( \N)ucQa{g]q$`;ZZڀQ-y_Bd&7_@S=Go'wGJNjP w֘p$ iRu8HtSa8 T,tP,:2a5镐n{^]qB..Cs (KG X2vLA&1̹$6bgט2 ]n$%y;ݮÔK uX3@uXEE7aBh OVbN] W 1a\d%G:JҌJNH ĉ۫8r$PU4|.t?Lo,oh݌|e_ktkD4*nM|7ȁ2,=/B;]'߾wws oΓD!EMc/>2ÐqU1x0]~SR(VF2Mp݀ 3Nī >o Af!884d‹deEr౤c\*7gF;ca43"| !xDA ” !gFbfo0?* &RRGݛ" _B\|0 "J;N+h# Z̏J,B{*ês}RaqzaN)MyE4VhKM{CSI0o;U={iFr$X ;hs/}{@ټ$?q6BOKS^$ز?N~n~% c`ȅ|!"/£ fvq QWq@2AR%Ya2Ȧ=(v:6({_l_oYOfOaDT0mZ0s'kO<" cF\O$ bۅZtqOģcFr9F[Fv dd|OtQSfRᓩfrӃjJd뱥c*ulg2Vfl:s 1oUK\jIjzE֠hUOX(­a9G`ڙ9TBsͣ^a386OC8H"dD敊Wh~5RDcW4@O%j&5'k,77 9[Vf}q85AOYluK2*I(%Vk0"V8@0ƒII"=`gl)w @fbn!el)1ƙWG<|XX:B0ؾc]wuVSbz<.j;n̽q#՗#rcr)UWWp\iֿgJME*.7`}6z0tѲZ5Ҭ؇O|h,qVlĤ‚wR@ŵ]{oL+9'aw2\MJYmS1/ 3V PݕuZc6Hs*>c7n* S΅4jbNril%FY*HCK ) ψظb'h#T|5Εx"RO 8ĐckbRh<7. H[A vm||%ǖ SZJE4^{ +)P4t#r-$^[/Li ak_iy]w8)&DEY/扑gѪI_һ`@!A УDP5;^7,x@ RS Σ 8Ҝpxû'txMđ1),*<:QO,09^/ Qr/3p;hXïً۟7;lRO N26uιdO;_yk"|0s ϫڊGٵx7'7>7”h)::1ssa L_Vah|]:辻Um WhxGtklL!nUW;|cbM yfMə R`og 6:U_^jBIz~tq@_mZ(Uyæ̟q.Ud9 NkI6HO-%iD3]}Ҽ In9b#<yJ%AFVdȂC42>.fea>eʞxA}#F }Nnj>}8;2퐻z-LS7^3.k&?qďTw}ўXy6N(QYX~}(8uS!{)-R&b$&bND*Ǟ-IT4yF_mHqnKHџ>$ ܛ􊨡EA P?cg_de0 =pVH Ho ظ@ixO} tKUdc$f *!r4:[i-p=0α> *7vvq6s:YBW 4#F"#d dbU\*(*JT|0ӯ Y~ ^Ղ0.t-wa}4e,i`7xG3f"h\+7MLxqhQ]@#;`˟BqJ:D*\)O(yg+%%yvX$όX(WYvAr.1AZt'\%>W^9+sFD1! yƹrs'ݵ.6V;bBqi񽒍| [R5 WMIϷG/٠AEIt?ݘvFz~cH\h-΃8⼬-d㋛a ycLH32I\y`EVIN!ڦ0lV`K١( tK*u'z9&d8ݍC5_Fzob!c>oO W{%hQ8NxK#<< #K&c>%K_ǯMVD%ZyjB;%cְbL>c[Fҋ酟wy҆`CZ#S`.e K.mh|Ƙ7NnWp HI]|,#)Y6`%B^]Ŧ-s?VuęJ1 p4q&[ɬTFJ3>C7Ir-,tV%΁\D.A8| qq@ŲOwt!A.σ\)xMo}qҫWg!9s?3BͭF)^03{6L&-h!`-H*Avfww4VX񬲡듖kg̠wу.8 !e ڙ0FM)Fg@|@ZZSX@=2$X U&86[-H$:MiVQE"jmaH%rA&h62@_nИeR&M΂@^!dd@F3PBFz0 OqkA``z I ԓpkJ[m8ySBo)*aJt'L֧g GVEk8k:Tu!#9c9f{ *!n _VA5EL=PBÐլ/ӿFA' ߺ=0# 槢=Pn 녫 6Sڴn>/__?濯EuK'!a}"_5#JB2 P%GyCHPC8jg[+̰%MWŨ-s<vY:2ER-h) "LwvypRܞ~& %=_/P5tN:){ڮ~wՕxḧZ=o5^PNwՏ>ִq!g#,'-wgOxm2y恾bGr'5xI48\XI/0?ۛj4ڧCt!@ƓѲY:H=ayg;VPJFn\tGpGWa 3 $y{JkJG[q%CNjर(a~'F0ZvDejH~—@d,s[EP~om;*kYNW-[bsQՇM#Y_D?a x1sW|l|1yEħ:- xQ}VTY&dtgnm T 2Cʫ]6GN,X++e71o͛!D}_׫s xq=`6_;2\ms Ͳs?97GQi뜋#1(9?K{.7h7>6g?  ֗c ׻m!@1?lib51aEKAhKnq{ڜߟj3.9bC\t -bwoyC @rKSMP2Yt 9|G\E)"h`q~6z< E lVM0x\ tѿꗜyxrr_ofM+PwaA/O/?7uI.Sv¾ M`\^ӟ ]J.L}g&=mCCVvuCEL=xӽ,Gvڭ.1Sw*ڭGwXYonDj&$䙋hRġ K=jĈN9hA&L/ݚg.{˔ڵ^`R9OnƷpU16~zus2~qyO"Dn^^Y|S X(ǡr% լ_A|=VڏMKk^tuE}H)|:{2_}h&^5L%,SFûK!S,^_=-a%IUJXO/k,x X"DdZL}RgآX;Qt;Q(vJN)[5`S1ѧjb XBߌ"> ڬjVBoܣ~| C W*9Z5j~cf]kc-%>~? ^t w{3lLݤ!nBًOITE, o$.* 2+[%&tj-Š(ɨA_n_!dJoraTZiS ᵨږ)A0J[-Hr=Du; _+hx_pmJs· Kiݸ+Ut);[O X.XpiI2$C)9ǡ)|XuRq2b7y/1qәp ag(]jQ'~vw?ۊОX՚;w Z4*&۴93h9Ӫt Qc-sAt)WF軟"o@s-?x5RqI{$`wgz7_86l,2% c%RAdyHQؖ(uUx&IHg4p+(8!D*K]"*S^l N̰d骢&&Mn {W e ՃG;^wyd1gO￧=Qࡼ7d~럣;ك켣a4=WLa1]w}BQ~*ۇ_Gmܯa1y{ƶ$@1Mg7;7 }7 kmys;O 0h"h~([qΙw=qR ‰ukk9H5 "Z(}5$hEДF#ܣ u *9¢و0b\#$ Ly4J~KbZ< +O[̑3 ?l2`23\_</Bi֝)eJ` %/Ⱦw! 5Jv:j?ćxlB4[5iI 5 V浣I%nh>"( ,ږJ Ő=tjGH4fr{UG@dȮ-zsmx94]_k"Q1%v67^H#׆VpsSŐϕTX٪TU0~8k>JC1)e QMӦ8~܀mca.(sV?59ʀf`> +"ژkfŁg<#Z8_97BrUSD:'S3G"Q_ Yljzl}bk Q>tu2Kۚ8҄7Mߐ69u9Io qB\jD5%%:ؐ3N8;k)T1/KY}|)U0taJ|<D$ompϕ^ dTBq0B;&4<1)PalGoC&m `ګՏl7%V^vF }Y00 P7[eJ*%L)cHPP@ck\*`('0/}n,U %U6EII5aId%H| Hdb;VhQz6Fkhg0;X!hƒ*9.; WOZJp"QM6jf62*瞠ґbC m =p5x@(xgɄAY΄ FN;jpvmٟŋ 2 @jsڽ}oQv T9~ Ԩ }g=T"Ξ$C^TA=LQ;(ҝl CFt>YûZ楰G . PZ`i4\( ~. e>H3tP(K#7Z168:8^[ƺM!0ļ6 5t u:X9!!@8VV}@C_a)w[F#LS #^5:/U ;ˮfT@/T:*[vE}>R)@xh91`k>!J`[T`kV"W2f4 Nl!pkhIJ5{oj 0f"щe%Yöu[Id-j𝖭'0e0CJxޔ̙^5Ä ځLje5=|PCU&!4w{Ia -77f]yydk ;b! kX,4UcNTSg8d꣇ŧd_`yoGv+[QK.MD4 码zќB\@}1s .,F1/M nDi}]T@*O'7׭/ɽЩCŜo/ځ#Ykpzb:Dրot!8 " =1L_qgB,㝗9jC5>fFhvܨ[Is9.08wxլm,#>!m(\mڭO;~Ԛ[SsP=woP2otp?ޖgQpmHSx.W|VJ\={k AJ!ɓEV!)e9w*0eBn)Bͭjue7j ǯSK ߙl~⦒Ł<-L9kƮN㽹w6g2Őѧ]i {H"Iω#'/I8I(bH "J ('D>$uwx>W)ʬ SN.M-9~ V`b<\IA Qo͉\oa$h*BK+ --pTbyΖPVvB< eg[_HVydu(Ą}i7IS8Z>7W 4TFQ9WtMBCJ}S=QPsVo~\iǁJ t7UoVΙ+P'siBF*Fxm紙$Sg}ԩ5I(sD*'Zf<V&pp\{y2 4._嬿\ySn/sh20ϡlzFHOE bx:j4W7|$q(hb>]̯/sU+\fxq5#;tyEYA7A *,^2uV"Dѝz=6bC!!?5佄_PW6@`MU)e=E*uRsOTAلH {_MvMI'inDQxa.gOu^S_p %7;V)*7N_ȓvdv:JTu޽oc4g|7Δmv.z T܄~ȥcO1|F<[ֳk59N_j 5>db4 OF$S?&~k@= s͛8Jx.t1n : ǽ| uO1t^9uڹf Z̍Iz=uڭVQ LHLԅ 1P DLk/\ A~=-Z- !4|ݟDJ;n0sPB{l̢ța"4#_Z|1LY/6Kf>cC)-?1e¶7F47B*LH| )b{#=ηaVG?^I`+2*]ǍϹ؞56%$n$έ8nbyŒdWs0Snx}y(<#?HS 2?]_TCpXyJ7^(P6`%)zfRCH$ά]z=kmPU*U+˂{ॗP*! 7pQՇhic+eqUj;#L*MZtVWAcNRGCݤMe㌮^i>cSFRb}Ж7< )]NJ#&jrUb n֟L QTf{ظLf|ܶfM J(Na aHQ8ŵ7w_PaX p+J2Ρ(B;[Hg },HZeHڥy{zS .Md$`@NFAGu0Dmk Ba3I)j8jD]\>~N-"S/CP뜣{Έם8Sx-m}PF4GÎ+(9ra9.yLo8 $n\I)'V BL=ڂ{@;Ȥia8 8h!+Dz^Ȩ]r9F+o5S(o @r[Xv\J^R DV*hc]2 e>uƥ%v1^R oG;}K4$Y5LW8XiBOlvsvS6V1v?u0(\m.9[ʕ%q1:sks/=Z#wey PF)[x `W|w._c=H[8$nd'{:^J ]xx tU0WaAU}Y,OZl+6Qs6ۉKJ^)1 d4:WvjJ#'K#G:sq4A43Ġ3QaIcƒH;o C8ocbϰweA0d4Eq4OvcfiAb̭>;bm,/'k9uk$ ?J%hLD㰂6њQ~;\#"_9ZpY)[ڊ*j4JpKMq]4_/p"%EF끮?8\,wі !@י圖F$ uرBB{{F*w:ZR VF|%+ڵ^K=اa2)*Ze1NiiHO\GjҜjTy):Nku+:w0Xmm +X@ ʓؽ/kl5dw˷A6[8"\y,e]]V03|hO\ȵŐJ` ^GR즳^WA6{8@ _v gJl]nk?Y<:w҂+ߩamjO淫go.l82Hye6Yg\0-_z5m~#ϋq2Y+u\\S얯3TZ픯p%3Kype\'. r~/E=Tp7n2qkO^.cwAl\%9KF=~yk<+[[+6p=$cnD?SGpM4Q]l pq{v6r@c-$6Il7dJ))(<(鈍\C严sj+ pfpf2vyW+܀-ME_S23q23R3_k 'Gzju;ltzzDG+*&qNXVgVTR^ q qր6^91'EXo5};0x'mT3 G\'~~Q{ÝHğvBD`j(Q%xxp>Dm! ]\>:gJgd)zI ж20`+n|ySWVu$J]; < |c:<އa1=;D7h#o<:mXģj1pcmյI}jz+McXbjrusi 4BPA= fNPrX,ԇAb5YMhn TPddT9bʖ^c~4f5XOٖ:ZL!!6xDUWF7և=<DSѝY{hbԾ˲mǑ&Jj[ÜFdY;0%7_b7bYdLpGY: }wA +gG]7?S5%ki”v^%A~ TD:~xdnXP3<ܰCvLw/EȻ4#xHQhvzX[2skS7tp%PPK7,nXڡ){&-ctQ@.eU;@QŐ8c1% NT&5%M6F֎5Ҏ/MgXUNbQ/m/dm)kem (Ue'_3{e_A gxlm }( a. Hyhl=a_=|a/9żp hc"r\l⧶ƚyaG](׳Π3_w{,.&B{s[<`ޫ?m,e?i:gv_.<{z3G.w='yy'ZaCh _ӳ|g [ "ۊYS]9ɺ1-*1BΨ.~^9qԞZBy=]{n?iW{‡:yalK/:{ayଷ3,Uw>|x|8cW? z[Ǝ|ׯiy;/ x`YL1hұ h@F!%iK PiL F9 ",PAP)^DŽe2  "8#eR2CNd3 [cIxtXH $ w-2PI:XM%PqPސTG,$N@rЩIa-xG"ݽ*^ڌ Rk? RIsQ|S͛7[1w>51\uRif(E(/JEyQ*.Jc4Mq6I+M#wPtU'ʫ:Q^Չ2 DUwuk\ "\rI Ob}QE9X_T%ѱM$_/76O Րo&*ik Og}C8Z,QQ"jBjَ>.`v~ec YS.*"DjkGӂiG—"y[=P0l؀N,{#4q?+IӸ sބ9HaΧnœ7aΛ0M&yuaq8e2F*&0Maafث|H!Q08Oe{ICbTHD*1z  Bӂх7 *oK-ͦ kӤ}%J<0Vؤp |s RQ-N%mMThe 〷9*[k5QJpPvmZP#kZÔc֒lkoSJ깙Y8P,af\C1k ^ͣ{4d%gB3F˫J7c 1iO?:c،1l6c 1Ս1,rģ Q(xTZbA3X%SR*oS >0aIy 5v6u0H˭uL'Ωc:hxh8ʾ o6JF}ﶡC*o}.!ǭzAj)ZG&NN솃ڏ{ ad|ar6`GqJBgUAWf_Jw/|zf_` IX$+$N29w/;'$$ZKiߟipitO=txŢ_Lʜdl 0oqt~GPAyoX [bLt 2LHq҄HZs|] lURIpFӠ/1˛Dޞjd~W-*r) eeO"7X8eBkc57y[-9Yre &z:͊rl}]tH7l8.$;qoA?XFyG2 頟B Lr#}ﯺQ*1&Psc_G*>}@3 3y$(! ,G J#*@1R꼣oԻEN dɵxt,t֜E;xk t(sƽz!Tڌ?<3=quʛ{=~m={]sT[ %om-ߢHejt7;a]䥸DyBV Kҥ*t!#`RxNUUkohԀ55Zy^yހ; f89C<*W5J\Вt⫌a!З 87\\͗+x9Q)p*\Mx,fFsgmsTQ]Ms7+.]I U>%VV5O[,9$ %g4#uDzGn5J~({4(4qzgлJ}, eEG~Wzuo5ORNvjO-տ–YNNm:$@>šv_ Wb`ܼib8;2qk9OuDwNРb' v~6-͑ Ɯ!j33Cnp AqXu;88x|{ÏPԐjMWuuZWuu6: f}`1h[.|H en مϧxH1z  㪘J-΢숂V: %6%`].z~yR纨/_{Bh0muX>&6--}n:r+~ʩi@rBmMd0lSW$~Nc=Ku%Gj{H;4w&qQ̖8/!|V녻4^vq< pB?!Nb8xt΀ƜQUK)l+0%?L;`Uqe)^|WQFnWQᶅ+pCpJkSBLpM?PVT?Qh]*Ͽ |P7GO,|ܛ{Z%[T[Žc)mgZ d-GR?vj7U}k?Z7믛n~E>܈UN tŕС֡ ms`JMo^n~!Owol-^bϾBMzIٸu.}Fuv-ih@K z=rvB5hK֛MΥZT+Mªౕ3 j=Pꕱ}bqH7'z͟]0@`z ekPd@$:|29qon}!2 ;G:ZKFsگYg`43ZںExSvF7qu?H?TrtyKIvms'hYV?ZފK?]vDoR?9„P,WoW;bMv֊6F$vf!e2P9 { [-zˤǛ2Eﬕ!-0vf<4eS4Oa¹ʹ'/:LR5~A uZ) }#x֛Dqb iͯ+o\=m~[ f$=QO'5󋏻7`d &Q5*j 51i0ދQ!/( >;J Hf -AԫW[zKPo -A a'ե]9~@=\ei]X ֥8P nn#z}y〽5wМ{`|lVpċnG ?$|CG]#ȶ :YsH3l o?"|B唂dHfFKQ"`p5U1l|trA3?lYK9џُKp .%p/pP'ԒoCaF\CDV9jKo"x>᤭=`7FxeX5\ k'Pzs u{P`ni/e5p:w%/:FMOGҾWwmXDUߦ%]T\TqbT2;0uHgk;CI՚|5֒*TUmV#H镯EsGxD۟/O.o '1\q䱥=}UN39P>\_GtF9t~$LFQ0@9s*k?59Fp6{Hb`/>CΪDĄ:S A3$V[Pszyq~-ʛvgZڞt\łՐ;RY4 }eAbȡؚo*ZjRIWJU)\1[bϓ M㼀n~L54 @yCe8"m,ۂI`}g8.:՘p!q(T E 6S'AFE6bf[!h5?d{X2 v]}9vyKyj9_n~/nG85 _fE,ۡ_W?~~nhg8$>kq{y^xNj^W7ʱ+&IvGJ#gNOVB̖ܶ:N^\lï˳K40|ވbUKRf-82''m[ SlX N)1g gƆ6sDH$LFQ3(&}9(ї6*fPd$&eR%nZ[a>hŚK`Z: *tji Uve SR^kœb ]HQ3Yh;0"鋲6+j0~,C{^u͞,Do$iպꑗ]VOtn(2Ɏz}&cTx'ojZԂߏx9.?~x*(B"cY8WXU0$5Ha&rUyNQ9DQܾM&b$)yi2ob&{X1uVe6fHN^m& SlfֈYE= *^1U_ckѧ('?Fwƨiy=b{8}E}md%9V@ؿ.+u;&Xg!cq %ĝu(E{Yz٥ ƜRHHBy6cqͯͦ{Lq>*?ߣH__xr"H y\DBol.S-<`IܦEMR1%|̢jL&Ԁ*_xpW-m3oZcL\U ۺ>oЈ*znC9x*-%GsA0gfߥO-U+vDZHXU[|T!Lțp-eA  Lg BmQ9CT.ws Yׁȉ2pG_Rg<G0aP)Ŏ(<S.AE9xf%tRuLY *MrjXҾj7AN*C#Ne7Z)m5Ϗ3Be WìG Z;Zg'Ry*grW-QF3k>;+IS#DnQ>`ږ#ٝVa,ΚClQN&d7 -8dB+PN-tx{IJ|a2 @S*x~zvW'8&ݟvy#.)]+ 5US Kd ASɅ(*bgkHj-FW(^8`I mbljͦ{^m-~~ٚ%Xm:ɸģ*ʶcp)T }.boCVBt\jFS_{H,("SfX>Bch)gڷY͝Sت<툈jS}R18_P@&XGJ LH,Ţk˶*eHcxAu#x0L{Ӎ$ZW[zˆ֖_Pߴ=`R=@t-Nn!ͺzHQ8l?,=փ1-`/ 8:LE=mT1z`7!WI@n-/>/Nɋ/MOV<bs:1lZ:&M)䜡jofôbC8[Ӹ4ԙ3~]vA8յf]kUUIAN51)rY\3BHۉ Zagލ⌐0kQI2g[G][ #R3vJ&}Xρi/^2rd(Cv8[TjI [X~ea]bJ ߗb+6U #XI>`UQs[f˾vo7n_^up0ALqiV:\D*4U´\SjjX1 m zUK _a-?1V#;-V.H-:|< \9vI.)]Z밲b)\lYUTf(jjEW R(f䓡ljgI{+ vxȇ ,; ɇb-n[d1}K>JRIb%]@ĶD|!#18lwJk.Sp͝SfЬKJDx?)PbPd Z,iE!09N8:.@ "o ץгfsEACnnP2 7pˠ( =.Z@~4 ]M=hl( ҽHPB Mہ>к7mӤ0}|,fyLGyI BH΅798Zh9ޡOδ.x :C7حۣH$VO}ĔPż}pEeWVeL%UQY 1+^)d膁(JK3m+RiQQuւ1C>d'T \>mm egb z $8T%20.rМ,9#(2]hAVEP(ż7DeMnkTsB&ӷym\v{F˵޾L~8v#sl˙w$2LB}ys%3ETC$˻kT{[Y0o"Yw|cp@j"Y AŽ'/ r/~ d9H^ NPh3Ï/A$"h gey``g{M* w/G<0PM.G5… JMiyϝprh/-";ߑ@o¦~6u,@3_~ *藫~y\Mo2I4@aCU(yn2ϕ*x:fR$9H!o_nӓN^t軋_*3qEhބӢb6:5MΛqb/a[׿t/b'áE&1^p4 3 zTpRW?Ic{V\I;wւU` EFaa6ʬ"e(I ʛlmMSvw8QLv+[G[mpف+V8F¥'jkbPX 3 <60S-Mkd( fJRxYkRsQnؔ-PmO3@Ypj>!T)/74[ʒxYKV9!rnJy #P#`S&T_fBKA/XQ`;a 6@^P?=` Ts /5Vj ضƁOY!UԿӘk;H.U!L.Bٰ"DJi6l"b:N)m^uPB:((Zp ybqr!Sܹl(`)ƶ Q{m m%%; ȸ>]`”nXc (AQP!J\閶ց2WJ6fһ? E)3jf_?>᪖<@Sw]܅߬1&ȴBe@Nɤğ6q"<"1]"hZK ⯮ oU_ {z㞚罹{4p Σo̷wpFBͼ{H7j.(_3?JXtGQ(o׀`'OwE `+*)C3B]>̊}5v/sFψy,Ne(Teٖ=l Nz: Yp\Jl> .?|mo(7O9 2S-rimQd[['gaD9LH6YQ(H&i=Hz*P_hʞ3^~hˆF!Wo0-/bH !U!h/^  ʵĦy4Nɽe^Q)$U&B0yDMaLր//ǛOn6'+WSvZχuA׾DF!uDdB )3%E^r&tYZRh<5eeͫץ*T^NNƃg[TJita~_۫.(%CIWd=Q(hgÓy{˼;Q 6w}nIUHj߸׷\<{x7/Xiw;{Ov\H)wJO NwmiVkf}kdl>.r>c%<|B4٫ f*Xf~1G/Xtwwz?O. ڇbJ:)rꌜAdBffV9vtU$?)Dʋ{1/%t~ ]':3]D:Pm|dڐ_O7x@p oo lD>Edl8#ԧ.2\AC=AJ{tp0$m;K0VEc**Yėo.]LUF"svݎ)eMĖ{KN' hvn8@ YTl~ ] (6~4ovu9iDhT~wX$ 3 gEiz{gO>ŦwZL:fZ\DeqބRpNhJ/2c@)rMhFKD˝wf];F$ d\񂴠ʐHxBEݗ-{+nJbB=ZU=UϟՈwH? ߨJ=Xj(mj ԡ+ d6E n5sWielA)4TKi Z[,P& );Fu`60xmDȌZw9-g!V#@UDjpS{ RҢ@6q6gG'iT^1p33cmV {ԳmqLP9,Jd):Q :df臟\2_[ο۽WJY3)E *X+\9/Kٖ=5Dޘ1S >巪(tψ\pP%vB΅LInN5a]СNFMd GU (?v,HzN(67\F8QٲJs(쬞9Rh:ׅwN&b^GMyQ*R9GyI AVzx 6HɄjh_lOן6.#QVRw'ÀQuv{VL9B#"Ȋڑq3"&+;gMz'/_=_-{I GXY? T:Z*=Vb/M·(4=oEωGٍnyrNb,\M@[EuPRv,UË1րmHB2wv`=YQ\E2{`$\66p@c<&%zZ(E 0B5Wև }D9Ry ,?b3k;[`Ҿh"Jp؇ZPҖ-S`8/ɫ#R֚s^W~Az' OLAv|.5xT j^بueU.D,Χ\"JrYYcZghH/2eg6bna~Av6Avq|ZldIeaCR(!Ģ=h Ϝs3pCA]FAm r FbܫTt\ +\~;UHGա H:D X'>Zfb#djj=5R Lպ Pn%5U9lv+i/.x:ݳ X S hWr4֘sOK}Y!e kXG }Z\:$# E(S1z)..m*cEd9HG ZDJ.խ.jݸ aP#W(+RXBstןvF,qR8э͹tLݽ_\ׅJw[^wo ӎFǡɐK(E`P%؋;eÀl [hѲ][HG3~:nPvPU`+$%RrGe}ʨdJpF;Z: ݃7Gz%\UhW.p-C^IьU0=V~&Ql?5?2|GMzOb LTL= {$fv9ak-9&8va]|NHaQwZGZLĢ /lH0 k $%ع[ QS"]L<[AKDl.Mv!8Gx!a} &0Ll!֧irO=5jqS1F H4s.oC z"脰 =)٭ 9>h1Z=Zoy? :$\ֻja)8L";[z S&z8!$Ybs҉J(CO׍>ulp$z%hi}^ʈ3E}gmKHy)ܾrA.֪[Z.ϥo*N}sT1A N*L&^";KV؀e3E` ;&U4ĉV@ Hc.;8XV%{.ۣSF% ,āT9b`pP_ xBײ^XRrlm-ʔb a h}Za΄P3 _LrӠMm;2ɃK~*HVBI~'(50ӧ,q5޵,t^4%eg#ZlP)*:aJMX*s9D Z YMxP/UO" 3^UK3a+/51#QR_! PGsrgnƸeԲj1 v1YbbQdD}")-4?25EgʿEcU_qDWF_CDRӖ>ѿYr Ȅ P-zS`^/{6ӡ4#AiA tngۄM7ȫZ,LvFD(;j"LQ&ZBd=Zr8n높+Y?iwx~Vh%BxvBxNz¸#yE0]o=xCۗflI-47"4λ,@=I]^ \`"19iF À Sg~uZgEN':N7bl>GЇƝvip(-7;!]Z}Z,:߸ !!AmZ:=$ﴖeUW1Fi.k@1B&c|CЏc1F@αd 3K#1- h1>!fh>142U286'g-箎\kNG3WR#p5Frs1!ms5:QgffPR';7g=Xk"~%W/ϋ]^<ןOg'oe<8)YQmB2,7ٴE8 #̙{8n ;]HI9^9x7̵Xtd~"]ёNjyѱsI$HsH'CZ<UM9z/pÙcX>ꌉ3c IqG).>.k{IY"VԚ:1i,eiL*skI+eGIU` q%u]Hz.iGS+ن` eC!`ZvO5ȖGhRlU.ۣhXbhTh'E ^*\1y?HV>^;;HtO7%nǀ;q*崣#u?F{J*U\ؕn?[pޥR6@Nc|QSj45sP@RrBR-b gLZ}}5|3e:hj8Gv''jj/ɉ“'ݹ 꽿X~'^6_Ϯ{, @JgM(\8Ft 4? [SݙWbL>k|4-Y7BVلUO@|ͼt --a)6\ɛ 8Q6RH@uk]`N{~&SקJ`ǹZ/=7uluLp`~8qX\?|zrk|sYI`'c'aT7sԃ^]}WՅz2/;nԲ~~Oron$O|܎,. LGg_{-f逋YMn^)@wz&jrZ 뎣􃛯/S?{_?׳:ݝ]yML#2r9t 9 ܛ2端)so)S(տb꫿clBd2)SʩzɦB|GQ7ᇫx^Yܼ/jܤM~|Ɨ)gܤk{̧_gJϓ*<;W+/B}~XՑ_RM7ay }+r~ZkwuƶN[2nŝ zR}8W?xqYpuϳ.S0>ҧ7הz1ϓ_J0 J|m|wR⻛gR~g/_~k5Ԓw9/^O4Iۮϒ!) @/b=>=Jln2 sԿYwI)evBWǤ(=Z"u _?/ɛ|!ܫejr6-Vo׊I.[Z77zL6WM͌îۧ3lƫeqv%;4wLMdl;癥y$QzݢD9!zE!Uxv݄a$g@OFnj40p,h*}T aAZq|pӓ}\w#'3 |Qtǃ[R/Ro>=\l|, a2XSO0^ pwmH_mFSޔk8{7Z4y+QH9ɥ_cА!1CIY 0}@w1Hx/뽃NbeNbgח>x^ߌzt)N{0$mr5`0;~Pٸ'K\9mx)Ͳ;S l<{}WF~_= 6n DlW<Q21"#Si+oiwPԖక~,Zw+.pV]x}39[ק&/_̈́)‰(!_3,S"pw chzK { EK u3K|w-m+ԡz|6őGjk-L+KNs66\qSԻ51/ӟz ^^ؓB _T^[(+"P"̑?]JiXQ$Jen?uahTZvF> drwp,bJ+aphj`ZcƉV!Jq$äs+AIwY9戃a2 $:~%6H) V I8\R͹sq9(=V(JRJi}YFڷ;r+ʚ;r+5w*r0aVaWi+ Ga4VXCUiጷ:"b0"bƊbl %# s=CvV= MO.3M_߿~Oz17nmOz1x2X(ALJP&%(J0bJ/w伝 35p%WlO{ N"8Np5YYYYgmhe#smeM濙ՠq4 Li'TP0M@Q:!Q"%`K%@$ ąynUipH,;G 362F簢?E9"PtC7ZUV"Td 1K{dItc8R{K fMPM(l0$dG̊,IIUOذ"Vnyjq0i'c x9,q q$x mp):7J@oeU R(bLcpXN1!pv|cPâB0røUam=m0end,LF2˦TksEOkaUx ]&]&]&]6yt2l̈;b\mJV^~wL)Hsch&9fԠ'[q7`3"LR<%(gSG 8,Mۇ.{CXAQ!v]X(ӎ2X(ӎc͈X<H(#H C()m2ZXF) LsR=-EE8qCb9ȬծUdF8, 'Q2E˴oEK!^Q!h u 3F=},[#[JIi=,roV)froV)f̽4  a ,.Ư{b9XB\6+F >o+ɭ(W[y{xTi˞i=? &@a9y.vL T%p2SI9/P/.M/ /4PvaB0?֮۬f8ĪOVR]wJ~߿m;qE9pbqql311.uUYⵤީ?N~AGw{7b-I9on 8LKFr%=-F[>z><&}(w/U^|Oo{"!Z6YN_e$˧~^*mki[l4;*y{L}宎6{ၲE^I-0YNRk;f9hqǙ 83u{8>z4Xų /\-ʗSko8ԧ}5^ϜgW,ˇh.O?V|u]ՋO wT.vSq:f29Nպ[gOj&(knp^P7"tI%T?T x Փ>}1; Rp5SmZ:"s}}7kee䶠\KY#ک_{憆-&^ϞXvS `#ӳh[zlc[6iJBGo^pg_Z7c<Ze3bO=R}-isMʜ5 :c^˷CZVD5N݅yb)A.obL"XT錴[GYz=4fhIw8'[x>s0Ԯk:Šjg:5#v=p3SBaRie:}IiMД,+oedsX4}5ZT˦tDׯU_y}?wKhtݹy|7?Oճ?>q:ڷE$hC6{eX)ͼ-qӂqB}̚eqG/~l:f~j&եB;s]ph$H߿{bW3ۣΜ8ƀ8R3=25,uX5UtMLսKҜ>ҙuh#YJ<ҙG:HgL>V?ҙG:HgVjiA1o=ӓ* j*RD Ƕ\N; IcOќ权2nv2򗧶/ͭ'$kq̀> ɵ!UC.b}Q;Ls9Lg$ ŇՇ+BUip!jiٻ涑$Wz1EVXvL8PkIg,";ԲHėYyUV%L^;u /6z_x՛P$o:hgTAM{Cv U+0]o YS>OAMzJҝ=ߌt 4 p^~N.=M/=MEE{CqsR?o3ilfP_~*ƼHYŏp/oqՏEzsSv o^Y zyN+=sso̷pdg;~v釽^pub;'k9}(s7ceA.$xpW$68\밺 ۇ&|?͍'~\T 볝R ɩ2]R"9@ iiwci?AҤl(n/eRvdO&ua~1@@ u*`& E É0I>_nelHLi)a]UVqW>BTHC 9T|2}ddκDT7o 6 g{*59k346Tg0 x:巏OU!M½NDUDGJie>RQDr.q.K9-)yqIqa,0aV//S_u,AD0RhKbBNm؎rw4HI.B6XQ9wnzYL/ u?4: d.=SH BWgA0淵#U#6qdgwC>˿w!s=ĢsgϮ.[A`]f X0ug#hiOZs{~b&嚚_f_d"WG*^;Ty;mL`IGJH18DIsi x 52+ NR0U!Ql b}4Lm@>DdynrxTJ ofT9qE,6qQ\!FWhei'+༈S/4H+C,7H' 9ᰣY5Mz/'{;tTg/ |*!]j*" %9'Z^48;=ml>Ǩ sPjנY3;-dqGLo-T%eLP7HC0B'J!voAXR9A(XKwF?Oqt:K߹1#u r %V0N8 ɷ;.UR6_tA5ek H3IKXH"X%-( &iY]'yRTb%WH"qKmKg"eRRK"HkQY]6(3^J6gasXsv93֟JPQ(QT8 !đ,c3cd {ʬU89lT ^ I"Š"тRY. 8BbT#-<$x .rp(0_ Lal2p_h€J#`u WՋ? lh-ɧs#+*C /zF$t`i9~'.L= 3NYG#D|9 jƎzz\IH!pqI Oax$x4(ʩnr,$Gt|}@,$:vY ZMt}_+/ C vO} q k3mHg~x𔟳1s >t/ntYjܝMd_S -Ou~YXEntCD)Wc!2Mp}9ȓ`'ffx1jw`"~|1l9[sK;.yӱn29f *':h/{.݌9Qf]G,?:cAfZrKoM7>BӺBRDtps8 mK; &cn NޭƅϴpvcGs༘$Z:NCΔٓX ebWҽ4~xeMlVpDYQwpTخEԦ< @`?l S'ݏYI$;Q*W(+Qf5HHW ;Â<g}Ri])''~A "]U ]^V1H̑ Gkڐ2w|6wݖEu-'IpffT yI Uma/IpȈm-1W)uO33G%!G[Y+UXS A:C:P(q+sm;ũ@M6q'Mԩ+ vYpb#J8KKʌ%-i0*G)ŝ8xF NǨUUB*21UnҬOI6MN-yLN85,lhr,6S|ߎsvF:̅G,Y6]%P7Ϧ7 >)Q%܏IsbWRo ԛIl4׏1WyEK#uhkf)9ǔGUf('ou$UɌ|M'&tGE[~ZJ+T Gt=7-\5&y sFlĝ(8G~3z_-뗧I*- ]dFUb놙pL<be38f nUT!~{^‰tAsI@6ƖYK"-C$HmU iA:+T$’0 ZB!*RiBh$$82T%|^)QDHb1gKff USdMg` lʰdm~HRX?$fm/t[B0Y5+q،jBc:n48#O9Rqޖ z o|(&>DeRfP;*8yZ\!낓$xnKڞլpRS?I˪K⯻PYˍO/ύ*{gkv$?zX9RxdJѪVm/ۅSNːQ HHS23qsBQ.1Cr`MbN;$b3\&X1E8YLsjqI)6ש @[' t$9Ŕ$D"B"g ydb#JcHDbV tY+Q2U=@,4Lj%L"![GH2I$%ցH2Mi¥>XH yQϥ0s cx0DZƸaτ<c^kzΓ|}`M84FJ(ˤ9u<\"N(MЄ.z$PCz_UZj+yWmVJAIO?枇9DVqCC܂I3k`q/ W DxK%~ȅvfe:h@U8P5|-66ЂOqt:KY$kw5 2t\lEqc-e8aR#9K)ɣ3l{j"ĞV0KWfxfr5K k =FKȥh l CP9vNҍ9cA֛Ky޸ av;g,ZrK$.yӱnC?f *':h/{ossW(ʶ9ok7^ǧɳ6˧R=SޗcΥ;`L=c\U̚lL'&zSI3p>_i1w:z\Ͻ_u'Y)(N=Vاf38->>?UZ+KcA_Sk-$K9h"NK%0o a&J\cVA"j\z͌ wj!Z4w/NG1&4|>U VTh(`VFx1\de"$R3&%+G!*.b~\F2)( kbѺ*HD[( !\6w%un:MB`u?:ȯM]\"]e(:.dP7ǫ4z7UȔZ[t7/|Ȥ:l '}B'"{~MfL:;Žm#IՁUnWWwWqٗ͗uUE)$ HIʑLptOOwOOԆS򽻸.'c{2y#Cr5wUH×'#,`RzA%c"s-=S#+lEPrӀJ*FbP2pbI㣃 ѧD&P/` `CoKxnfnB*X#vmb1?rs#r?.L&81` d)y-G2\up.^l6J"JpRQI(pFX`Xn>ec3YwBrӻ8]/*U!NPdFy:5ZC9Q @gCMAҡN `ש5#K,V >XXR4Qg0ESj)!ǐ8Y00˹dw- \|d)QtZSr$[]PtCN(9TLe+][r^B(ꀹw%Y0!a["VmwQL`RKHpNR]*KgR[M*BB#bc?*1 $ ~Y[;$˄VUqc ō522*Jx ޓ*s4|XVҊ1bѠ`c.ӶfxfC AS~(%poZA&˴PV]GZ^Mgʗ+a+7wWwx507GA+ѵ9mAxr: trkۈ %o,ax@ :bXl< WE'eȠVoI:㇄<=ӧ|>ނ|mu)P|7ZƟrVś+0J sؐIBlKhZ( )-ydȬys]ș&<\y,w󚄻uXH+4jg)]q7PREw#T5XhJ*aH#.ވ'/Vs<f(Ƶ|sE7V|PH;ᮏܪXczeIo qU9SMxE%?oR+^Guߢm)C׵>{\ dCe ļ2i٧-?a,o F$X̔URB 1.r;d㯹 EOr}-I^' .Kj7VaS8 ;<:lE=;ӕ(E$P!&9ϧ+*{?w@3]ec!6aلb6?-9ŗPMΓ- ʛhl\^B:@wKz %+`[lJO盰cJS*{<6Cdn v"t dUP~M9KetwZ  !^K4tx&,N{jSa,eg=҄@T?%0׌#t;ђWY뒤g_{ZՕ`d4Ta?/o;>r//%໗eٯHb-1&,uޓb޶R՝Mq*͡(:g6#kD`S&eZF6C $*p'BgNIԊPPIT~#R d;w[7O"B:潬br0Gڥ<_{Ji|ve.}ˎ$#%eBүiHA0 hSNU!1 s j)-FX6iG']xT&@kɩl\.0T Ƥ]M;o )$ nJYkbkřk&f l併PU9O34n6_̻ʰfBf+b0}T*5)Ŕ SDaPD0&Zw`zpp(yidX{> W_ဥG ꓄xō&L?FD<-(Lηp{^.6#yE0{C|8 @J귒*l7ǯdCaq,&Sp,>Z^}c|@[1>@"%ik!^Uc7_Hk&W`@IHC KR!p{F_ss"~I2|4{A6Ue4ߕaec#<[7yQ.C鑫 SmE7,@# 3 G쇻ߦq*"̖dZm>C[O?*e&735\[fOQ_dlee屃2K{_1͘7}eM_yT#cL[}|pf}&qdVV.??8@;{Tte<;"'nZįqYԐ^\Wn`ݵAYi7vCuc`hgWi8@Tn08R(PED"8eHr)dBy!> @* oEQ9r[g@s(q?ۻ_#}DuTbBdzT~Dfs26~|?dbO!R(ʥ5ZxfW^Qi+-T{!GqsbxC)./!g@U)/ߺ5SsJvs ,*9+ePV0E6_sl#Or}r[SruJn.'Fp]s(R*o .\oW8.ΖK- Nt}*jg tB1?-9ŗP1lcysxc}:] N|)TsCe2 Q$:P9Å0*|t|t|ǘ(jV=& #JF>s(.<#:Rmrd̙0dXca]Rڤ )#hɔ`0(,BV2:Dv9RQ&4! q)=K' LIc0r9S:8[xlM NOXT iJ-{FX&DGE@NJOX(z['(գЬ]Lp~c3]GgH;(1Ez ("Q GilD;XVBl#pDL4cͼn<6`/RĸlQ xkRX %^/E6I0)~Nz0Y L ]\jf\M}F]^42{V[rEK64+*+vMNu1Չe*fb͘kJk(9lD(.,`*e'iH躥ϻ+b>d"Q0 IpF#0Fh ӫl|y-ox oQ+c$4"(z.MEiuBB6Gδl,G *6[X(8|cX ^]+-v{9@F:)r2"@3gRh襓뜑1cx`~}١]0p9࡚`v+4:Ą}0;u}bȡδD-0Gho3V$RdX>1ewJ:!R'G!n#WquG6^Im&u:4(5n=p[gqlXIiN3Xlsl-Bd{@CYzΆќnqb7->Ln{c|$6#oyl3?d$JTb$c콐@U^-Oc\f|p*1sv4%*Yo`=0N/ƎgAPGJnVK:!Wcy:2܍CnzdXMqݦn]ֽp-ƽsakhfEdvO+9G6Ji =3=ϘB`r^scMp1#{wiKL}OrG |n"CXRʦ_ʗtUxm<مS‘Ƞ!PR9>hJɣ>)-J#ā(WqKK]-{C%'Y]X4hf)8rnhR|ozTӭ#`v[5oRS6~::lf3qx^F jaL{cX_/_o~N,eq$o] y* D>HgW '~MZ|vWsT$gv)2t{{/9 YS1vP1MXQ8Rwm:PhV nA_o9_:ۑ,0f8R-1[as*[I#kYHƧ_?>7io7w39l:+|L b9hVeSDΈ:,8a9'hJވz`YgrXpVŐoO*rWg )zH?u[|# +8sߢYV=SlZ+ kx=Ãrֽ]Owx #8oEHF~^EDRV|O?08B9% yÿ2nז˗|lgafg]/~4L;O>t[ <b-E9=︄GR v*K˘w6 zJe͚'ݜ0x#z-N?&Yu8Vv@(+6LDZ#9JN%e̗mbjMi._ )P~ sz~F$`SԻJј^&|Lk=nj9&t(IT'B&ᣅ!*MAA (C?܏/ۡHKtK/:7 lcQRB猝Winlax@ B- s*^C S3: PbU6tDr!>YHKl$ԩR[K=ـ1hZUO,3,Fdm;rv!wy `h1ȧVazkg N]ªdŎ|Oj֍OX7bFՇkY7Řu+uMI w^]gǁά$aќvꨖ~^zRf(i"$pH@38 3_b<Awd\ ‹BӉ=J}T'W'ʭ\ BE7F$Ԡuݤ.-RӀ\ Phv)H @ IWY(-l 8g97}:^ZT-O2!5RF;(yڨ)lϛ 7HgL : oerTS.߾W4}gS^b3F}3<tEblV8Ds:Mu,dͤo9  275 DgՈ^>IBw7ALEdi{!˸} 1ݚ+U׉ 9GwN4;)r9q>Gt?ף9ymU3ߐ_oo?ߑwovb~K Z ~J m3byIc dJTݭۨ !j^*i3; &eIz zmG@@/lN5~ݮɨk N6#Cx;} N?Bʐ:z)R.0 ݑw<""NMH^SZdK`;WLT*\L(vi3\fla`aŸLASYs@x~1nv² +J mũ+ :񞢮[ox-WB#XT˱zLv5cY}|̆0 V!bnm܄xYCm$I9 yi v*:'O>g&Ԫ Z+[Crk-Z__'"*|!G$|9-d82~]Ţݥ0#-bi-ɏjot+r==%m=xfmiTf=zW5Wa17Q' i08{󏭮]jjMsF]]}dX[ 53_v{JGU7PpžT{L++' v_&RMJpEAvXm(8Rg;L3or ((df a]d1L$kY\{A Tۇ'кZ˹]_GU|.Vw7]⺫R\w5,='W6ʅYe륖?SImEM6m N3]ܚNzlMwSǑFʤ1CP-J0{98</7 ; :#ÃG9UY=-~2cq%SEjsd;ǃwkJbk[Wu СդeˊK5UX6yI0hHLZGI|t(4篦Y;WXM TKՈ ,[&,WU$t${AlqdMk2[&DO: 41xJćB?JRۧcQqZDZʻk/^"9 2ٳ.:llluKYiJWTB~APCcxhIX;3q`+E{^DL"/D! ZIS?8Hy |=f Ψ%c)fy Z15Xx/1hfA>ul d WAkAq^[>giiߎ twP '| pOI1cQe>$?&yzI?d;찥%@y8Pd4Fx=qv<GUy\Ģ,Fr6K9b1MR }&v 7s@[`#Oyy]p C7za_IkTdSo ډɦh}l-粠Ύ<'%%hזo29|3^*ٓҊ-t\O //~L%{Ǿjf;0{H4|Jќ Ge:za˪Ӫ%}x N"RsY)_-։,Vi '%aIySDTbcUTl g#=:,\i C $af)*}^| " r)k 'B6YY:ۜu4G% E ݔ>/hFҍmMSSt _1|%miZ @: :zcfO{}))XyQ*+ONF<^TNh%15"idkԀA0yȹB=v0oBK{syd\8p._Xj¹0}`Sj]}0mT݆5aĉC$)8N {Yg= \}/pUÜeWTADI1l6(} !m&^RW)'k}IyP'Hxjpёj8P{j8}pY^axdOぇZP8IS=п}9Ow?sm|yvߜtw_o$ԅp#w(-r[vO_~)_uq#"K _o&/܀&mmdI4sIG4KK&[rhVX,* ƷNf4^.K#yŧξwS3o^w[z:y6beX{N_=N?OR䯙;Q%uNS9I]vGlݍ @j8K0B+c4)di_[P[͠BZ%ϹˆHREAiPB-6&9NI.ծuVqbREB {pίa61΁ [޹u>A8Q[' `D:Lijn/4uj-撣d|3IiI Lea3Έ$ӚLX;Gs,G;OJǰrP"%GJGN4R(!B FRZ8_F1ÛThFD/ $X[8ow>K%H2s.P2:Őr4ܓqQMHF'~3ÌRr3wsfV yI:>ʯbBH\t檟dV+}3P,8MO0@ %@ic a6:~^.Ah׼!缯 76\˻xV+ jghL^My=VSsmÇa 6>< U&JNa׭{'37`S|s: q@}!y'`\k)L},4VkcXx) B%6I Fu rI6+ 9P,|(^_.rT1}(W#[S)ߖW#7o藷TN?H]* ;H `Xt9?O8(=W|kV!UV_ȃ:2A@5W΁>l[/Y]x^Pr!)91%pJSa  `%NIvgq%cy!Լ!A9wUQL哱&52 y%3Ywh5Yfj4U4_2}q6'Yb>}-Wsit̟LE@"%bQe) Ln jpN/ k:O6_ӛLaLA2*k# JBף JejHHoxk'>Iߵ4_ ܼCI鑀t_) 킨G{5@)GSA^bJO݄QP6S[6ܖ>EKC;(6Azmҿ!ݾ;k@Bؔal[Csu]M/+nZi;w'fP%>34\e7y`{](Pѝ^g}.C Fݠ;]b;Ōx5J;﹖f3x[=}@d@$:?=ۙ"AD~#Apm GUy߾{vxkߍч?G]DG8gSW8 ibu^ "8ub$ .9b.SػLōЈ2k"P:;QhẮ рЂct8ƅErO\jMY;콏 E!V QU*_ |bE G^mFcpɭR4PZn5 :[X3;,Y'ܼ☧r  %4iq"06Hɑoomt(EM6cSw|?˲$큈 Z%tN*2;4&8Bj5VzHKJ(2V!HiC t*RRi${na:O!YɐZq#LVREȻV 4ZD'j-Rf E؃z1%Ds1+ħ/1 h.;5o;"d\eQ {3UFBүGʣ#hbՏSĮc[ADBڳGI!LOЭ2 3y \dИ pUFR[ AK(5JˊeQ&'3X3,Œp(3 IrCLf5xW};\yь f4Km'rt e:a"gX<f(Wλ? "akHcA6IrI0g2pFA %?JBJˡ]̟̯:ӣ@+cܣo@~r$ٵdߖ1x" HծKfsxbYrivu%Εӳk K!DPu&E5R>-qz$;茳sɝ[sv N/pdxvkv8>[rn6!ؕvC=nNb] 1֜٫};S6p^Y 0gӳ2@fxWqE7߂+Vk[ W'LwQޝJ^פг G}_vݹ+3HZ,2_秋m(ShYL*3b,ɴf(V(kï!WFBX?FDoѓ!w\0kӷRj 75CR1q] RoB(faOWZinݻ~#UL}8>5WP""(!5bJ,'9Nh*4 z͑Y ZvywI#{;[M_髧MF5!!dR_][NsOx!Wܞ1nu3ysyڠe5d)yTmui: czP[4ۛ`Osݺw>.F&;K_Ѻ)3uJF˫v}_XJ 3ir1AH4Vq#õkZZp)!ctn/g*h_# &/9 +ze?r{&;w#$`D h3 U< -!ˡ뷱enaVTCA˙p2-UN(DR04#~ly9I.zs4'8(iڹc`g{jƉN9Jo;(Vuokˍє[hP#ɀ)xQX Fk)F#b5 &|" kݻ#B:L (%ÈKڷ*y6yލEry?I1MϿmi X\fMAYY-b7ayP)k+7r7+/jzXFVi*n3Dˬvw洑%Qo~Ԡ G!i[+r;YE]TU\9%\3^'`lPnV̎&㼻OtT+`" 1a'!퀽p1"vXK)#!nt!v'JQZ-\7SD8"K>8? `+x3[7Y>Oi]}*"ҊyEuO`w|~Dwٝ{-2ņQvU0wl9DeeE;ZqyYa]C1Dh9eGk1@sjH}Dq< EMO2O y&bu7 (5x=?d!C8G<_}7Ypl؀\O  w8zxhZeL`n&cп|D1\NدwwW=tF6Uek%M%u*bL3d֥F6`fFL|4G, A&ag~aעf7 9^rTډ7aVvi/+7CD[>;mM.G?mI[l#gt$kFhzhn:L՟]V*X+G<^F`%aθh6 kɋ}h v՛dlb;^uՕDt2pɚ,Bm Mzᓐ$-cuVV)ų+6Gf%3 n_vNVB1*ޔgpI0u&9]&j " eBD: mx+h7o{E{}1sMPcrMkj{ã1o'9G~˗'(l9#co5W9pZ<&?/p}`BWLVw~款-P)gĵV?qmS̀i oή Oy5{ػ6 _L憞o_a F@훳?,|! n|VYiX?(\ 5_ǽ9FN73:]Wڌ-[>rHӕZ2uT1Pr N^m8?Yl XdѬUUiק=D|Kz\l16AV;BeүG@E:>h nYЃ>"Jd4L Vμr?n>`sD a11&Je Xv 1?) F #w95B +TRU4^)$w-3F>2$XK~}Bg'͖1 \jbX@% Aw^Pl!3눍gBT_zeY HTو'ݍPb"4N*|#Ѣ;Nx80(ƄlBZ!`k !3zu\o>,d@)yLfY5cgs!2OT<̶KEhE/@ޗʡY$H9-4mq!1i=w±h ΄⠘qg6ywޝ}2_]:+Ϗm iens@=h:c6l)r2vŸ3fI<|hG6O{n"WBJBX(FoMpYF14eeSm5b9]lɢW5N  V3d:hEk+RHDZT+v^izNӪ12Br+575Y2b0(}@iAm|ȥ,!Ætxl,> ֵPG儑.ڐ%-Z'Kg1F#'H$8LJl>1J-m'GMMy|NPus]GWwGv0\˾J&$ҭ[|\dxV])8\5-Ӓ?>t;i!~۝TFxɟ>OR-֊,"K/Xo"@ Et7}؆R1A!kP wYǙ#t^Psh0hUdsiи L,ep$j)#*k/ 4Oу5}3C(}BR-6_ ˮaG|9*J6di/يR̍YP$q˼tK%]ACA\Q%z”CdQaL{zl{cvNk?엖u엖~i_4.i0L f -*fgr`fxk QٮؕS}{PNFE(Ŧlm نG'Go C5p`@v7Jy}ҳK/@grP7Ō^.2`PA1!h&l1gӓd ճhyg=&=RMܑfű_4yVy`/ų[fNϥ1E|級gU$bt6 ^X0k:~~꥛UrMٗOy,h1S|/!NHӝYnF,\Em&c3*c@Nh6OO61=Z vSI1HG:*K:aI1EWcLPqY"QL5-+ UA 2.T.8ieL7Zyes=tOoEA~+W787x~|4z[Lj{I̓W6H9U}r^82gQ+ٹڴNM&|V-%OW+9/g_y5&)~y#+蕭j+IG+VBkmd8±޿Yo1OjMCq~/,ܞqI;gwJ>J!rWU)Βc[|~]^*#..Yz#""ZP (ʁ~ =ZZ[p1ƀeK**#X:Βk:lEߴNdA'K'U)zA>3`uեWٗd^-Uw%a\>Y!daKO[xR@S? umA:_kH5I62$Q<_>P#glR@ _: ØJ\yh2&ĩm B} ,yk  J)f@ Z_6vxABU:5o%ۇEt-q2KȮO^8}Պ\D՗\s٪ېzq h:/?\ @3$==N]}Hwi!ne=vCeOSQkzV {9 vn@؄?OmĤ =6gέu_X<bYi$|4O[J=F粭y/Q 3QՐ gj}s#o!uIXCAO Rf2gW ~ -w|0Yw>uLTO/:hMpRO+'*(ZՎ͟G-c;EˍF+CK!P(?Z^ym Gg,Pq`q6:.N0`!'&h%j-(x۸00$['$ Z-J g=xS`k~,PQhCw 4+9H -ŰhEc,/cy6퀓Yp2GVvH M4f4@z|'/R>TF_;ۙZڕx3os&fD5Z"T"z'Dl cT' 9N[^SY=ƥ ]4wT2a`2K&JG&+-CʹC\@ Հ]; eEx8 sI2/Aǀ5'HLI{aH\y9&ZMbܲI7_fe4W[PF--zL✲vK=&iy=&`0fJv\@i{\M/%G#yӥr0*< M+z֬}5|0ɖKn>gvh[b1?-"-N& }J xߖ}$Gp#tjP+"\5y]2 "T[Xo[da6bn˳~tdi5k(rM$>.NP;ZI>umUW7.a֚Y-r?>(Ic*/Vu`U0HÒ4IW[XS}]]2cuxh' S9yhh./.axqY,oGE"-fcҶ,Td&iXiqq-c9(5,+Qi|vR4H`\]. ]67ժoȓՅZ(&Fnݭ͗(D*k$52R6%?9l 8Ȧ/oEWxy<΃eۏ5ZГ>N( PG#A{ d1eRU E>k( }@bX?U˖1̔A.A 1sq^cf|Πlf%;1+P\vNcY 7|CgZP3?Q6ȗ7I_ LVUd|JprjkcɊ%9ߢd==؃ &"ŧm\:9'7[ }|u!虢SRIGQ%29-Rҁ=`>emßT@F-|ԱKN߮~+Ir[u<%_:eE>K*Ehf'#Dݕ)i6j#QKQxP;`l0ʠza@APZ@hb|KRA8!|")XTٹuZ(c60b:Ԕ06g2vsme8slc4 HkNDեEf UuA[\]<յ;MDFtau,$ZZcbS?S霛Xn8@6K G(,dR_RCŃ׊t.:bL yL8(JIU8^ U0b&6O280"VPJV r@H9PlrA)(0.2˻A!س*TR2#ݔ&UwCBiܜ&<О%5;TH%w40K*O]D?2DB#WwեRz@ ‡)m Q;zzT泋|qOתݟ,Xkd|\^; 7r4 ";unJ^Ȼyt6&4%VmBwpdv{NcpdakcIw30~gt!K_ 6d485'|IGPPeGa|&76 _P\rq1l'J^1>ɦ!;PQF0IdJ1(:sGW ER:ԸXe֣N*8Q2!zDsz㗡wej-+ 3- pYaglk[LݨӖol6+fj>d&il8&<z44b;H'aϴ;6db6.].|6^/'[-0tdqAʅ!Np&dr-7,hIc*f_%Y%d;ld%K@1GB;%E`'fEː&dɐpZ̾La9􀧭IW4U!+D&rlDF*HVciv,EK=dLݭ81$ @*{,Eىn;VQ#-xR D\AT@B!;J`PZ4&e b Jkq* Ὼ|s|u~Y߮w{g3aoC*?1 n@3C1^sZf:@tͥxR;xBUjl)u,#g.-CqΘjlΣrƟ}rG1)mI) a/34{lqG,c8q:cihv1g9[k$l'k<:k^N~FA?Cdzv5UG-^'Fp{+ͦF2k\ /Q<$1CX7DpZJo'#`nSXzAc(,& !9caْ#+3y~ ee拆YlHjpɂ4\koHV T޼i Ml7Vu[:89:'e`0޼ʘtUbigO񊌤V_$l mIM,%b呔9 xg8 XjeͶKݸn z.4A^-{nR Gz{H~(ѵ-/2:REg41J%DVA0\ 0G0%+PA:%g9D tЌxSx%z 1Nu78K +ʡ$m1g)HVzcP[L{dՄm=z6]R+Y- D\ )I,Nڋ(4`:`cU:4݇!sV0g5sV0gJ$3Dыr ɚlE:F?bae%kaX(P_Ώ"k8ڜ%-/НՀtgÀn{""&e} @(6 E!DҳZ Z(v TI)ƴ6PגKځrM]46Vrps>jAuQ@qFoo퉩&|p!G%H}%;v{F#B~OC>(ngz{zKs0bz#Q {[@vDp(f\_#NxM[`Z1D%υg08pм`{!!Q;qPM'gd"Pv'w`|aÇ,Bfp={?y$$,=7cUhtvfs*$4}/Lc C9!6ƚj5y48@VLԩZcG$Frd,ZpxРeVQ 9XS 0aKm,H[z$vԑV/[kfgCk棧Cl8,6Զe٠2D} H]BA9HgPGJQrD(Rb2'ϿTEZCmZ!%RrE!I dE%$]T2R\ȀZ T]6ek$R`đv:*A [P "(uP&)L9N(shuMHO«Sxݧ+Qjք%`ѧHc"ڂ 2hǮXru9wϮkE8}%d\}$2Yr!bѣRb ^RzK/b+FTdc4IågkWwx|n@{vJ+Sg}x$ĜtkN^/$g6@D+RRخjv,/5jYb8@[mYek4ʚew4S߲XYY:KFG}_]{] a$Ak-qђC'&O\`r%T,qc|Plp05_y߬\7;^_ne:RE]~q.D?Ua_\_`]S A"j;M2ƢOP6 .gW[-vn:X1DPJ7H鋐&*„:YnV5O:]j,?ޝޞHaeݗria`ߝGV7 C!-BbޒtxxCSX 6{#)Lki) ?^2L,/?~I48L>腡00Vy(9נ̴/a~s e"-8nAIW{yvK"'6FÃyyOjTP 8n)J|Nvq{} !ݐO?bO>hI &@qEGOw8y;Omi~r_)&kGihtF:˵CKJ%`Jm)K:?~u׵ekcG/|{CZ-2dKhhk oX3fo98w|h}m'hj[5]iyxHcDHF=Rpt޻07c"*uh٢< {5,x@ƁJ$o̮&`$$D̠vxAU^)]؃:$}dꭲ/W;yK6# pΠJ,Ϭ Wl~%+wwsz[s|߿sKx~ȯ\/ۛN*Z ޒkmVE0.rN C@mpHMkZT6C$djO. C_߄YQ agO:%ܮ/0ۄn/uiϧ@.,o 8sK9|h;<CKe*A 0!l.aBd"oH买Lv=M5sԆjtAy(@W"䔋 +I/.!2-u3-t,c{Ȯ<"Zj㍏?h!^zx>Z#-{'SXQyxx^['˨xk*ܩ VO-Kz>X.B\Ҵ +[U\~Ks-zަMB<)k/9bIwrD^.Y#hTj"IK=7D}Ѽtf1QX rCֻ8P {1K]9ypPP:~V}EyF!'ضQpBQLc_M[}I$8iuZcH"F3=X_JcEDD2R۞?fdjd"PindB# u{P5^uWp.U-#R G8^3n*jX$m@!SL\)ڗHKLI B)`ĴIc",K4 5bVɾ)B=LQNU4%$"qTʼnR/TrHLLȸYa؄Kܺ*Ld5K9K#X3儲0CVJI QJ/ap^, 1h70ec#A(rMFIRpQTJ,3* c! PuU'8=vkkDg[BMtg ܺt]6afTIv&zg;c Ctۙ]+%j$Ȓ=O'"vT>I„dc1+v6ƇjihܶTt R0D!l1;s]fGVHs02f]`vLhBMB05*mv x7>28"s.ܓu -bb|TJe) t2J@ƫM_ݓQ&5eT*]o .e,)9o$~( n +A:Rg'r! )?Ӕ*NoQZXjMQ ^x=S#ZW<_!x$QEյcU;633NI IIITP[!w0HD"agN4_ǮxG)\%x*{>>I)IG7}qK-u<}E1yܡ=hmRFD; uF\0Jx: .x+ɶڽ+-=h?ژ_{~w}K%.A ~&C 5w ֯*" C!}?+y8R5Mp!p`8}> h,jMW, "0Q_0_V}9S}u(Gk(7Ai͢F(#;p_`x]J8PY^K˺n`;l>c׿iapk#3ɓҼpӂɿŵ…/Yo;58ݾ\ fw߽7&_ ⺻pkCv|j lzm5KwuJ~'ᅛL ^5{gk,+a'+ ʋ2u˵طkXwe_]#+'ڽ .7$lVAlYޛ2XVW.$'. Ep=s?nݿ?q+~7ܽ?|p?8{?/ӳ>z(ZA߿;{1h0×f~~o_$Y?> e}a7]ǘ`@~ Sr- F R&o?ƙw66l<ğ{ӟ`&3@Ab @vE8 #@{Vt\V*aޫMaf}1ۈ]t~ngV &PNϷv{/M^goPXr5=LOիg`\6Ql?|o츨qyes]Z\-?.V?ݶ8 l qd Zx^5d>UΧ$:8/cLu澯үe =fww|NWa&em=OL@fMŮ3'`>d>ɣGΠxWVVoڟY2UǏG%;{uzUs6< ,_\B| 8nK QpJ \Y~ "`&m&¹}pn_(  '02$QBАmF1T\!HC8M8\{(IP*XgW۞ܻЗ^nc_$+G;U+u޻o!ցh$_k:;)Tz \f,%#*PxUNY2CL1Sj@PViDBEH&FAqF) P)bz[Ѯ *it^TyS.AvzpܽyP<(ǽg+2%5Q"S"-u45:Z=g"L,&Ev΢xV휮3#}нy2,;^{C-D6wļIsZ%Գ9Wa_R\HJh\Ht-1d>‡վZBIaj"j,)IQB;j@8 e)WޡCBsy)og!vs; zwnSYrP(:ogIy?,dqXuHܩGzՀ|}F< D^ Mh3sݺyAs[2wc~O*'zRr]2Vy^xe-e!PPƞMi؍]Tfn!IPkMڀČt)BV,]rLIaIɷ8EXRJu|;ꉏ5ȼyB!ƯZMY f>^EС ng 6=}yܕ H&2 +>Hl<<+rY j-9}"ϥ{K 9Z9N'% "[8`by]W{:'qHIj:%bĶ=7bCy߼p'Uxk!ʙx,aCgf3se-͕U7c;rҞFw5n{n}2i$+ں"7вeHw7;=z*$T uX*|<ߴ5ha̝զܟLԴX$(H]=Yu/M"Q_nL_vYknG˨*(&^VoAvՕQ\Pf+~48 9aL6q,$HD |es% ~Bj*Nu9O,-hFYɦRdS⪶BA T!$4B2' MQ(LS*ɥL JhBC#*h*"& r4R|8 @SrŖ9K6yeTm$cY@n= \6aݙOy,p}ZΆ:{"j$[Zi^:^]V”6*trr,aer(VH$9HI&kc11Iͦk:-gy#Y T* U*BA@foY:6Ҧ70iez)6S!&裯[kDvNM" &)"'>: `[jU ): ~gkft4^ko8/lZ[Zo !j e9#\eC0]$QUFi p )ZRdfEh-Yb)a6բZn;īb<}AߩO6=O=w0WvGq=_7u8ri3.AW%`z m)A 7^c8h$PP0/Le>sR#Iv c]( 3ܞ6 R؈R#QU>S]~e"o"TE6FaW-3F1D[^6ZJ(Ql˷ #&'@~)5 utEiP8ᚄɑQc'tvQ92O54s"Kf9re m9Z*po3UsNICjQMHܼiT yV6%*Vsɔx6=8(~r>|pٚf\J@ s6*9p(_WʣISrnw2q% Ū1ΓKSss:Y뙟zvNxwEKf57àS%{c^Xu#[?4ApBn/Qʃ%{Cj8ed s_J:/zq_鏦 C~mD0LiC6 ,[^x@9dC L(aIAS 2zR^5Vzv?Mثtye#P$e+mw^Pڻ욧+h`_޷Tc =mGA_5R/_{c}EZGs*ۃ;'[jj[\r}N'ˋ{8-jǩ P˛=/=}&g[+~uqTyh֬W.z Kـ`-xV_QM}W6!td[Ty{rp|O?[b_NR( MD3&F%5!?}z@w $dJB0TYaɀ,U?]4@&@7v~H'AL.o<{?~8Ϸu?tN?Z7C'{]]<Ȁ-TeO%:K'xݑI]xN./OY! I:85`Jqn-!!80tKQ₂[qP%ִ PgDw[9zxlnjz*ڋ:gPm)3/FQ$'ȥ:3nY٢˭g7}|fQ_ NB@v.k^|xǐ(Ɛ%sJJwZ8lk=YGZݜZ`DvY?v:0 {duЗn|UIihQ8R]QGn @ϋ/%Q݁=:I*zx쎕D|ӈC*}nȮn ]S 8%>܁\Cd&W2UC6(˝Rzu+ܔ^ ]qIףXೠ hs<9]C Lnz89-UDN8;^?:]SUNJM.N#ǸXTű0i冄HQ qKu8zΏ㵄j{H7q*if,)d ILJYo2+"5Ę n"!S&pS@NN0*_vg ,oXI!\\6茩i/W"#&X@f)#ĔkȒxP[!(awlUF@)l!1 ;}' y;h]K4BK\1LľhoZ8Q$u0nht`ٕEf&YRk2h OI w@_~уvqRIRѭJBn[0;1n@?鰞~6MVY,?\7Ւs&հ—R9Wrtݯ(_o. PkB3iYR,O8w(DfirO\l@g͕s'E6աPXܳ"?aң}ЊIax&z!,m?:>e<}ŏg. Dzm5:\y.֏pv}vJ .cW ,ʍOa_$t7KuHl,ۅDo#Gul; D>~4\Yv:QXM)iL٢ ;)%MI+tz[պjle'Oȩ/Z3ʼnT}S8Y}vcdomɅfmo {+ԀNNN8u_izN>-Å/Y,T"]61H\bM/oz_!9b:7]8!{~I~Բ(i sZ'`ɛȌTˈ 7Wܨ@Ze$ -ʓ|Um\HGA4Į y%or&rhƥk'i;Z3fIXM <;)CsE0ъAiƓLX($:0 Wls%+"=Sɗ%<4/.3aY#3Ǝw hMYzL(һNK$]GAU|`Hi 7< -Ұx}Up\s2ɾgؗ1XRb4 cow (XY7C[Ib7Kih! ;ǎhO^ҥ:rw E;l:]-NAHN*t&::#0Ҝ';ŋ+KU[R$U=dRz:dy|%,n.w Emɳ:3TӊD%AJ|k@|U)wmeVG-aI !3fz}ӓ-^9s#Z⥜ I'qZ^55("YJ7t+^bP$ 3}w87$N\ @zX95^4,)}w R.|1aTqdɀS<"y^)f4J_Nz>bc8;Cc g)![\Vfm\hՐA,;oaёxdצ:|O8xj*_!!ۙkВHO^^D*\qIDfێ6$`%_ ?c+pS+:jKlċ՗8@ԣB]nTsB玀_ZU#ig:jQ DMAG8gުsʐ7f>)>A ۽q>TKTF󅱫qm2.tІZBU6/.ݶZ5\EYIN$vqζf2deȕ!bސWHNjщ<g󹎄}!-L؜BJN(>sbokToKl˗rvΛ}2N)w}87Z3ߑ1BS?W aJy23(P27=+SZFCr(pHFO[>m`Ozirp:s8;qn:u[ǡEO~[/'N8;͉ݼȿ7 (s&ӑf TdyP\h<-FeY:*hז&8yGK͗[fu3bl4vOw{A:vUJ/CUdu[y3r0$?\To/wpM4y6y6m ZVnBDcd|nK]B;>•+g&o99QB )S,)hGF=~w( ѷi=U w^/.øϺ9tL&2N'ߝR5n:ut@E1ode;1agزE-,ALw ٹNqku|wwz#884,]Ӻ>Bc Pcwz#]s6#bnu~1r:.*.ԁIlej=P0`4\1Z/.y!Ba1Y?E=4&-{1@(PO;3 NI=OaAӠ,[n?yс.#a醧U{ p!3MSX|`٦cYd~/uodfl#PW.4~|PhQ2]ྸh[w v;d>|V" mo:IӠ-}LJw^wC ˃SMsj,u uEiw=郵jSZ?^͢I)ffe5h +Φg7x`ƒtފ5 VqNUoD=ȦŋP| 3|N_"QRzxxJ% WCF j%a}MjjnrM̠vnj4زG PWaÿ_pf7~Tʼnuw?V;Kedq1A[dSDt ;jbpq+zUVcOxULh;$&d!/ 3PViOG R5~~+E ZNjO 춶"@ =Yŋr ?HK*YjstɺK~0'kdRK.sV@Z*ܺ {8[ ;k#`kTg2vGr=Sg+H~nw9@X { 7Z_?ު>wcwd ulP@ީ}X+;<;b=@|+vb&o. =n 1ǪN, 3EV"ǠJCF68EAm}C0:hdIx_8iLkz Dhʠf>Kr L^R&1C2eYyScN,^L|&Xzyi(!K 4x=4d6wKꓫ}fc >{( KNf`ϤbD,2DDɒ1hl`j1%xIr`hTd"h,6,Q)ruN(h H"C!d-GKJq Ǹ%t-Iћ\-/x^˿=O \^_2sqNYWDW3jW:ɫųlCaӪ[$dY?ǩuW2ǩ /Ыs̩E'qMdUi!D)-21JԛfqW#5TRc0qݥ3kn8I R@{`HnWi 5/|bh΂v'Cw'n9D>c/ 8ŵGy~l0^'#f? ӌZ68*f]Ը *if^Y[HC-Cvo5|\ٟ&tve)QCi{_~+l2gIwg qfM&hgO9XΕ9FeMQxio}>qz2M}z=0H&i^lS?-̟9%-oF_S)qŴҦOO$`TJX]0V?ܧA^椖\dU$!qaBEy6%C1hs:q\) zyE$(vvn1=W ⿞ "(FS(B$63+ñer }V j,4r[TbVq-ٖ{Gmu /__p mAZ$0N !XCv9;t6ґ&cLD<$ݨVk;9(}D$'т,|.hr`:.= qƨ1g1  bdIo*9.vXYsE%o.Y*x!R蹼)Jv,-T2ɵZꄇS$*sH0 9@ F%D0G%P#IFAZR >H?*4 <R7*Z i*n5;z-uGY{g"m-;zIu҂{X/}#HyR:E af\&("SNX< ֊i<7O7uJ(ު[|\jNwXg.: Lk\Bdg.{㚛*CqxЈ;5jnY,=/zGr8cdGqMQ++"7h|~SdbxStG7sۃܴ9 sTqnAu'#n.Z[}_ǀ|?! Z8CG<\CG:ݰzJAAbTFLZ8:d"6)Q:0Vy@ "Pg/H-Z g5>_^(Jjԯj$7O[ 5\5ߦK5r*)D6p1HRq&6o~>RCx|s75R-7 +@&+rtP؞wSXe~-3"J?0K1חK_z6_(=yu͗!BheUٌ^>v  `C+`c$)m~S^I O'[F>$@ߤnSĆBA;g0+Aj3,/Vl3&G0*55//tWOHlk<4$_h"9/OWɓ-swnjn#=~RvܜϚ0/WW:`bH!r~.6q!e"%Tc-;H1Z57ʌ"TbP t_ّЋ3~|*YV BŎ Qid_:_-w4|ikH^ +M 3Y&Ez]V$^s(5?(T%^6jSUb><6^KaZ!xKrccl K{+Y60/:% =xIUB/(0:sQ K$ۑ ZiKqnh}0:F!U<ˆ<ؙ 3CV> ЇfF l3PJٳp %#8nE9k,zi[Ɂ=" gLjI֗#\4\+Sf֥Ɗt!\3#GƬ1  U=JeybHe ҅kplcO֛&0]"y#=F, #ٻ洑W(sξfR^ۛ6l'[[KW DG s|‡8 FRtϥs!3ԁܘH3Nq*QnAu n[A@ ZNrEŚK֗I8ލe$|Q΂z] /*G׳^pO2pct͠1 Z:-ºѡ5Ѓzlj<:1VYo <h(h4\ypwu~StCչ1qqf%]) *[q` - ӝLsk97ׇ߳^ZC.Eʝyp;^2AXGl)[/OW!xl-SoCl]xt}1OlM./ mݺfx[>7Svxko3O_qƈ1 ոsNEZ`|ſ.و|41ÜS!k\WYZL匥_^=[|laȬnyI\.::̨SKܬN~ɋ+nsZm3v*SzylaB^lkH[(dB&gHj77>h[ Tgcv̑>=C4w1WT*&d:CP SK((T3SR* ]tN448י5]]Rd;T;u˖La$B:&c S5]-nImW rwpbDSQMWs._DdF%FqAVJywu9Y!0_~܇:bURs,o$/m;:Vd-+5;sC-x_b5oe ]l/\fl"k.X*cr M'rhː?$sI_/*oqCۻפ<|#4QI4ZmncE "J0*̓˧S51osMiL%ש/+XKyJ¼#XDFʚgڅ܆ݕy-l$lW{}L0cG7eDˊwN I"u f744<{ijl5ew{> z߹*7npV`?uIybѪS Nׄ 7ƃF?gUߤ`n80GQ Aj'nL;DɻT/f$,:X 7FFYrDzj0]7^Z LdW'Qm3G7 VÑnziwF| ϶,/聲J]o͕]?4u]y$2 ”Dy:GAƺZ#m5Ә yQb}q:A^փuffcڵc&>y*Mۃ{6^؀ɛbHM.0 lL{q- !(G\mjѰ5KQG4Xzllf5=Ѧi‚nbf#ױզ^d(+$B58m 졣 QIݏSwSq[l/:f[7TzhV&g¦իXW-rڙ kEߪgRWJMRUDR-jU uS2!`J<, m jO]vP )) qIF\,bl8SW0p m7H:iìMA>jC4+?}:4f0<> la[:ϧgjhrܡ d`5[}܋Q#Ci[ha:!& :6 _AiC~lMnx(*TQ/SaD&_\"pàD7o:W]臞o9jq0H0u?Ms40 ،Sdնe͓~K Gzگxn? ;>7_ni^<&?#\9/_/ܽ=yیΡwz ߑC}j11K3%[`H >5 —ό IC)!}]6+uf$T';;;x<fvw\\_8O_;ۓ[7v]5˫?Nqg|xur'Џ±p?M0ӵu ]u&i~{*TE-,\? R³zO=aA໘9'`=?t(uGӜLDztuj.zd[7|7%na,>GfA~\x ;tUTQFOsÓiu zS]Ժ�< 0ݻNnorƵND{s!uWx :XOmm1ɝaox B Nߎȉns߆iivan`p>7#$Q% ?OtFMz,,&DU1QUߟe*2wp5OEIiYsc2 Aվ| YgiOA%FB>UNst9,a70^d6:O1w'8n6_|K(g#H*h%> 0?5e1 ,lUO>jl}:2}œΦ`NT!" nPG94P孀S<dg+<A![ovt#OJK8>RG;98?gajYv$Z`}|Ly?i&ER#$&rfy~DnϼW\7`"'ܟ)*ܟyPbHЀ #O1pRg?˷WׁbZjA SDբc;>ÈZ4o/"J'\#>:>,>ah)r,iXynk j#iZG06|8Ef 20v2C R#Ch{?e!\Ď(띡qCqsz?u2y)Fu= iն;:Z֖SV`p/66ЌN܇(Ξ҂+S)Q0}v76~Ѩ9֘Ftӄ9|%HbYO[mg01-XB-@x 2S^I@9g@mQ@^G ԀQD^Zh5)0=iN7lc!]%ӈ P75)tGN=d%![Dt4@c5iЇOoJ Н E9& [pqdn&\4uu!!,Re+Ms7o#dPڊwr˃|f^ҬH.Q&xSꧽu[ǡMnʖC|7,V]g}pLgaͤ/m. +9Z8Q[c]ň L Zj$y{ {4x;ؠ%lۆj, e[Guیl먷˄VKL{9aVCt.Y s\[ zc柤.įv}d'xfgfi5.z&om5}_ZQţf#M׉4rѓ5_:=LdyڔLkG1=Sy%G5{~آf}lU-FlmZZJl죧[q9 M9öJCaQ=%eSTJb#uJ V8^ru$-!IHU mgeF/بSZ\]I֒%2TOV[\r.M:u4:7J׊+)\Rr P9L`TV8]Yfef/-l R+g`~J V>-HP9y ȍBe^AK֭3Dzd8<OHnKNc [[(4C 0?P*nkUzjTp>]eL382b]S!8v wnJ]Y0(2!L (v)wpiC&щpA2n_u};{.0\*$ݏ5|y**\jiM`[Fp5?)gJ GjYp)A gl 6Gj6]`9J x1<=֜;=?j8@#SSWȉr^~qrB:DZ4sT9) +l!'gg~ $Oottby>]g@$qkg1Qu,ɾ]?ʃ8OMF wC4(#gN>ܸ`! L3A>NE"^hȶ".rQK .D5HEn̜Mx)͇A>ftD5MUtY\U:񴙱WVrT`zp" 0cx"S8!Rv2qc0q\oJ@b* /5,%.vcE-HW%L.K1pUoy&_ ="ݕyΑWvAAa +OU.>y-ʭ{\i6>W{< 9/ 0ƒ'|J<_Զ|E /Hr{Qc2em(ωw@qiL}Ȃ= OjP˯?P \*Z*^WRΰ ʼnoۿDQHSɗT%9]쐃mM6F9~{z򬎐t~ROlYSՑF~g4v2k#5xR^qDeW7]:MduG]38sF-;CQx4DWn;7[?t~?p:q.$eo@oܗOE(5=I|91$+$+_Μt!!]tTR305I [RænxreX QAI.=*7<= @,i!ۢ MB Hm3[b(ȹYE5ܨg"dIHLQɔLl񊉠OHG2q.)kf*J)_? BVG¼G5K/!Le;bUZ^;m[i>1m/t޴Nxxm64.HYeɄċcHH.],P b!<Gi"&ʭzPs]9yBD߯-8H'%j_NE hN$dX;,0) J~$5SY'KnaES0 4 g)s 'c۴ſ1Mtè Ǻ q#^zJC;3+o鳝:Vqwp!!M2X4n̆THA"F,pl a` :燭If˕MP?ⱼO,VUvFyΓomtĊdCP"KYY$潙|0=/XD05~Y6ź 7~݀ n"DB׎!29 IQ3 4Bg<7m=! ] Rm~q(B==YٞDT?xFfwWMD kê w>Ga5*5S"}X#R{]v**ރjQܯ7˗}7HaO;F3϶2('=.毶SOfRyUT*5ԩ+*ZĐ2Yå6g`E8$OǵUb#[Az  S-"M-RGRӈRA(8BQLpL*`H)HGvm29FU@x:L'[qMK;aUv*peN3|f=D#Vs81*K9Rjrڝb}Z4H/XJI [ 0q bL>=8=0L 3?v:5;NFS%~3 G h ;" 2 I2jxZ éD'g\>FZhB0>M|Q>t9LXr) 7 r8yü,M +Wh\ 7&"PrpIG^\jAh ␀0*3O"~0I")#"8xbjTD4A:A*:hkWrBRU_x", P.}8':BM"h! ,qpILJjPv.o0`Cs.&0 c RS#kV8 T&fuDm_9Aqa.xT<ךB(ZK١kVe_]J]!{K[X`Zj$ccݞl\qpf_%nn\cl7a͊a= h(72ŗ|k{}mۄ3Ֆw9_d-f{ג:K[g^źDYa.^a]JK^rWT{jg]JԢs^DO{hwHYv,fnsd3][^8t(d/$o󔮟tm[qƀiETu,;_VnԬ16GViHz{>"bxa-_+Vy)Ժ={w噕s9.g&.$MF^h^u~`T!ctUΌ[7 K0#qKn9i=0!D[nu񃢭<+dhBFm7ٴbn co;>2ٞ`al.Y޶W-xSJwܷҝy4kYu^$(}i35Kj!4e׸7@w6푋gvߩ.<7ݙ(B4ޔF.EKuzwBWACp4ɪ˲=MZ kWooB*ӄCzLԁbh ltIegV]Q*F.i.V ICկ!ZM(80i?W(tj ɾ(X׮BZ͓ 84x0󪦼tqbheV5wMTG0} T@#L -Bzrdž‹}:2eas?r4CΞn'N!qCz+l,x0z b;/;F:ׇ:PAB"Y3@% ƅ2;ٞvx_;Iц@/+ 3vǣ~Üv1P<4p5\vP~̳X?Ru9VԮsqh$N`a/40d&~z_ xv㋾m0{܍a(J"9C5#,6.T땸\5UDӽ?|:,3#-g\j @'A28 ꏣ˴vN? ~a_fr3vnM'qp}{Go ff}Ӷh[Ys#y f_.tT>M{Bo(^ ple)r|vo7)Y}1 mjεE> P?)ba}znVB7ta(\03N3W";ѱ@XQ'FmsGS7fn$Խ(\LT EƅXPcETDrVP{9Hqagm (& ܓ˗#igx2ɧѴ8dN\o tcb@k+wBi7Rs)ޭ ݕZy/"Ғf4Ϊ^/%UCѕ(tؒ_A;"$jH M] _SV࿀?~ q5ٱe> ( lbxm-:`ۗ/6(/ɠb^ +r'\g}3YU:o4n9]F0=.&m8uZt2{wB* nPj<Ǣʻ(+_:*n9꨼4X<Qfm T@6]Aldq}s7lI"µM|V?7}j]-$d2vWS`ބO;a7@SVv8iR7ҕe_3N,<|&i |"u! ?LZhMwi s \^#h]f'dA ј5j#Caٌc#ɪLl5m?۩=8ɶZF26ͤ9O FxKO_5GI\2J~g98p?!4e0ڞZx[VCx2P 3wť#@\yxiLqGԺ Z$׬$1[D5|NV'<9hRE݈M7aV?C÷-Aq]C6s猣&!)!o‡lpI@8xi[9PUr"ˏz+&m!Kl>ıgbS'tb1qWtS 0Dyxeܭz.)?>r0尯RY7dPI䤊 0bx'cgde~V |F1y[Sqi R?݅YILrjw/t{-8$Urz}4 $4Zz> nj <ӾQZooDGF~D#7&XC^sTc§kc&adN1@xa{6Asc;24K/m ~sCk3xCy©=!ҟp2nlPDs" k![O5ϫ_@M[5`?z87晉?}Tw\VRq) ճ B+yFE{H"@#9T" |!\EPqFRsYBDC3+Eb6$ ;[ޖeR)Kʬ*Uh$GPbN[ȁQl81lNHZ~ū}7pitͅm4x"#<ԩT}0P0Q#z =@$[ 9jWD%pU_4\LUY̦y6Bpd%x,Lt1mTE-zja_2nNlfSG^.AkTnqTIw,EwCt+~}QĠܶSۿ>sh48ׯ8:7<5߱,G[TZi{[jK7pmkǕ.~:鰼tAN׆>a#ă<`gۏe޹ɘkp2(mN:k΅Kkf״3u56b9wz28MeKOw.)Wʚ ʇz#%Ŕ"`Zdc`CYx>=C¥k %\p.C875E1s4}Б#G:*0n/>|l,C20yO_Xh|?bV_7ya£*8*kdT3]d6 I`}u c& DIk$ RI1N3ߨ|ps WUug)wV޺8s>e_\JZ_VB NO*< .H*s|tolzN;m*zƟx;޻㣫{Rj/ֳbUmPzNXh a]y +>\|JyԱ΅h(CNDձ!͘J6L&(c!@vJT!Ad7ABMGK'JY6V-QH 6j1h1jĦ$O1Db-'Db#WOyRMWu#(>߻ំ?+tV:GEY~g~+SꩉXF_ME'*clyq?Q)񼢓Qo&3GqO(%Y<2bM*1@z~K:n-Zm N>M슖 UÅ-TŲ- ;*Ɩ(h:R($Ln{ti2-ٳMta 8h5c^jA~A8{[FzPImO!['M#vQ+:?eW( tN^M%v ;!RR5BB!ڐo]L 5SVhKWFp(uMs=ukGw];떏wM<>q~ӣ!;VtW\V+uҐ_m4I.Wm\$!ME*^Է_/j$5qιOc;{˵S"yD[I*G<]6}2 1~0l9&" <=  sr)2qR@E[+/+rg];h!kgMOaic_*g^ϥ[}YйQbVK n&U-YTPudL-іCUFĭV{Sp],JQiJ<m\QoCօJ\cu%L^ѶfA\Eaf^ {5R 3)3 ۭۻ)IUjӎF}) {0gmmꝩ'a BDO9H\ODA9 -R͂}x'Z u=¾vKs;g=$:07X008ğ{13 Qo{U8 7uzԾogWuSogXɏBY8?jn?EO1^֫6s>>Gwg%O[i+ Z;q#{mvzB*UCv9|k]csmvP)@2$ۺH2aZ\ ijm\3g;gJ]]ĥC&VW1SzK4Z"[S<.k%[DqCCSP4 zjSjFgle6Vh-ߏtqc^4;3[,@aU>`ABcm#wFRts'oi.- b]UITWEr X C#!F)CZ-J'Մ unsî3}}K0$ִFh ݸq4jrkz!4> jGG-*n°/vnċ|眵C $Bg*aU|ḻ%cB5a62I/V`ďtd*v* dU46HU/ eœJaLm.!TVYA4h%VY 9K DBN0H&NOcܱ },ˣI lلFUԠwkSuI|L!z z8[o0w .rL!Xz`5AHB9c^L'q2{o4Ui*=m$2m+6BW{O^~(; YSFk+{?lGlJ @-C(8 @-),[ K=(pb>LrE3x%ƈYbCa^um$BkOcDr37JaǑ}oNX"mvrV7{/ʡF!,b ‰V,kq8.1gQh O&[=z=G UZKANDDsI'gLS7Ryسx5 5l< WBjF ސnњ7Voط()^fz4M ܘn`.n#MAMvE0^Y ~^d4y&"q,C+h⣸bmOю`;P3jVsS;#5C˻TK9m$^oFH#HҵIxHxr{x Lx#؆MS,iۼ_ڟѶn9,xt vډ-7ۣ;w&0[}YIOI(L|1QND-OCJU!ms3%qL㭕lKY~>^ޜ]Gl@uE}C8SHHjHq*2ܵ."@gW'.[@YwW_9eE^L˴1Vꑒr@A#PLEcޕ5#nLI=tXv~pmx":P,n[*$y!P 3MtMl&,ii]4u1_U#Rp{ҫL;ƀ⻒Trc#ͮ -V)2 n/Ka(f\$XS: TU"vܧw :NIĈQ !6 4FƒĜRhBU(J;o :p4\V cQJʾR ;R,ekԾﰽ縰T:b nd*!f`*泵~flxk~eēy)ɶ<~]uB4ikk k.ְ]r$H]N Э f%UBCUw9*ϱ[>aV'B`;o7hX=NW]>~0W6/jh$=]c<huzNo{%ćabLC۬İ$%Yw? ?nzPwߟ?P#-AS4?, {4ή6K 2+~}oQU#:y iNovRȾ+ޭ%21ƒȳ,Lxc9"I'߂dqxN3Gia=pagۼ;؍{$q9![a5zuŭq B9-';~%FMOG]O`buH>y;=^Ԛ93-ctXTH-k=dKz^9Ozw 'O 1]9\ػ}u=y_(!]cNwsE:qx+8\ >Xc4n^F;뼆 8\iQ.TVv:#:{uE# >lVM҉ޥ0h06AoRS.|6?zj3c@s'mթ'h+:{jϾ@Wnָl5d9RfN;-\E+rWQ+mb2WߔtAC~umrSIrO=|ɀtZ \%DضB "vG\!\[EQTn?FZ7E-Vi2L/^m ʓ~IwKanPoQ}>J)q6@ N9tLG GRpH#ZHbu !K0,,!A1m ۅK Zkⴼk),\l\% ЃȾR_gWyr2 t ZiN:T(넘XSKkiq"#%UVc]+ZUuPr%avכ$*IHG86K< q%{4 J,"(Uam'u$b5\S̫=#8$]ivƛbAb[FX! 7J1 J TDa47痑j@jfcL8dY4%(INX%V`nN<dW2 F!ufއW.QŕsWzs0Y^q({Ԝxc.\ ]Z7~I`,M.9b|{g0W2S)RycG޸GB1Pt72LӄSi1 ` ,Ni4'18 eJ(iL&ɦJnwKD-^V_?>=wổ㟻v~2%2*cs)pQ啜s{|Kؕc5|_2wމ;0%A{%kdM'9 C>Y\P룋J֩[mp$JP/:妮p^1cT*qY%=.&} ZN(7kݬxK4 Qߐ;>XqhJ ,qD?O[52=>١ˇ000"a-%:W6THtnO%\ Bt Je"wNZDwW|,xuYD'_+!,33p8Qء&^voZ ýr@| y^z:ZWp c|&ycWzKq;?b0~x˜ӌ^82vdv429d̟nnѸ'bY!W>):az? 3k4^-: ~Q3Og] dۑ>r9H8ޘ]zܬai}䣏 㦕oL)'j!Hݍ"Pn h8$I{']ܭj8<HQH Cmg.f=P8$$+)kȾ呡 f:NK:48c 2Ir}sle^9%!wb.nR4py-:°&=wi?^BF,P^pE+tEmm *ED%d%k'4){y[lʈ7ɚcof@8oNz_@x]q,VP;OY}?_ܓH5pSe,IH!ElF` v]rgz@DhʙNx@ȔF/ł":8s0ڦ*( "bMƁզ>aTnQZj^p))#Բ*Pt_-OS+YF,^2c<:⑭g|~~aCBiȷ 5WʧPyF*B ;IpJS$Pan뮻9\>]A!쮝>;4ȓꕤ20tEU׈8Tڷ5NSF)is~*xg 4*z'y t)O8ORHv V}|b/efby8I /.Z >;8m$xӝ>ƫ:m'*#sk}:X\w^AM5g3F%kwՙJJw. '@X NsC9031}%$: ^G"0*̊*}Y 9Ol͌ WXvºaܭiA؝RiA - 2_"," 3UAR*۪TT`5zA_9bR܀P$a^&fc1{]/@.CT2#J ~2YZPK/y5P &ek ܵWX)St1DO? N)8ק"RyBlKJ`|%Jx.c| f(3$<+Hpt Wx^9oQܞ|#s)4rn? |He ^r`iW#}8(+٥UkQ3VXb9+7\ KF! -y1~~~φLhĘj/ŏ:\IK21 S9=7g͍63 RinmL4圤IdJc~$X%)@bs%ߕFGmlUHG8Mu%g&:h$i¹tr{e&aDK5SJŮ\J(=>geb U@9Q2X@PM(uC<D(9<%&+Sڇ+%T2J(IEJ2^YgH``S=B NzҎפQ[Gq˿uWP1+ cyX.yxߨ<^R '=atWo qs/MF-ݩ-ﶍ~3衊X$aSaE E6XjCJ4J %aDK]euSu tF!ǯql>mUe [!*А`EhH+%$Ė$'ZEyFӈ(ҜZ$uPfmwe0fgi|kȧS%ѕXy;AA\zLՒ0zT ($VyqgꈒrgI:*Lñ1HGFo2*$4yYv0$#kK HRK&LEa2$"RE NE!eQ4BZG2:駥b j =nWJe8lt,ӁfXɏV1rjX/v|wC犳aUm&&n:klqω*P9A] bR@le'Rޚ}$REFW t ٤3#, ܓeBȼd$:_~ŹzRg ky}7-Y]K\P"aCÛfѓ3a]iEI T :(ҕ)ay`G|ynz+ aD[@wGp(S ҫë7@  ă(R ;b"jN;s_"ٴK*8ġׇ 4mn#z廇 ݂ӌ*aӸܥpq=~z fbD1" " a"dlM6vA4z<6c14[Ͽ,޼f`i`߿%oo`0 9~MNah ,g/7@di^f% )Y= H\*E.O6 ǟ}K9P9GDWf>̙hde'V=mnoe=bh|`KG-(peL#<$lzWjv!5h|&[Vƞc 5솅@@y[kb\` ۱1"h2Ay g;LLx!E, ԡq @.`X3AD". 6|(<^Jߟ!sN85@E8n&DhW9#Ak=pNdxZMs3 ΥcիV)lTKb!]C?[-qꈟ2hוH\4W-9-w ⩽JT=RPPޙrgH*u)PJqAyhNX'\D:wIt]o!3,.Iv?0ױ6Hg+JG$4E "K]/U#kZZ_/F;m9Ԙ40 ^3gtljܻwT8 tjy~Txspۥ,i+U|>gs5^uؾ靚ug*H/IV0![ؠھh"nHϚ E]mE@66>ܭlZbAYK^{;(L :"b0z 'R*M6-+6cSk cNFU-$vͪzGmT>Dҕ&QNRn>XM'f {QqUoԷIYkuq :2yCs+Nm&Lg֏K (~ pK;[](W3>IRΤ2V)\ lgŗ5٦`^me khJB!V*yINѲжMJ \w^x'R8uT>ܒӡ{Fh-Ӭ0d,>M Y%^%eѽDv!uo[UɀSrF|3ޢS 0;Pɴۖד~]VLTEV.3 Td,r@ #JZ565Pb9B7N-??`Eg"7ow#Bf_Q;cn/\'rwd&?/`N7cdvΦa2ޮ{xdY}ӫM';ɍa?v v)7~.X>}Ꮟ ~w\O.Ex`J@Awたkb. \>\_,0?I_exU„(c|;8tvx3 ǟt?'YMg[­;~RuѬ0ʊ9o\IuIR NP_q\P0cMF ғza=KeC2ptArG <)-GlB s, Ժ"aR :H6Xs+9HK|pްr6,}UeeYdђ48LrɰˆBm@6X~'H2{ID41]&@\ΒI0VZնF u3LuX=-S4A9EKƨڞ5O4fMSfى}:.UdD4<5^*6A.;WWJ"L[dIt!x۶q3k6gW6-T0 Ff:z}\(fz.Ifer- .j{6M'axePi?G]}zl wggKJ.~x.X퓣{֍C.mU_-(@m=u4 G}އvKfmz饠@ek71l`K~7WFKm4a ޼IӥT@7Ӑ[5TAqBD]um1*hT^EުXRm,(fz/M_}j>B nfix),ψ;<+ 4㧯{ɏ4m,0ӍnX&^.IՔm|{9FP*R 6M5F6޼s2J d|fq퐂VG_8o NO?})bA[ 6[wVeرSe - @={ w͖R/vt՚)Q؂"SkԾJ#U"X-$=' RwH"l<} ^3gtlXvxjtْ?{۸#_ =~ȇC;-vd[0觭Ė=< pUK~Pi7$EM&@<3rUuuuu=[ yk&rzܤ {J1ySJ@0ȥQǙS`X '"8"ߕhMn ոvmtf1D1D;֌ڨ9<5k!D20%IVQʐ?N qJ9菎+T_ӿ0Z@HU@:v1vw퀝3 }rR4&1`Ԭ32xWZ."p ::Na}ʄO]@'ٸ;5E0 7"[Zz5U-˻Ǘw]&HVd"X1K-s2YZߪd!9CQ[Wܚ 2Rژc\OR_:f&Z]"Bgwʳ n.#?1?$~?ͅYWtyH3ep,d0֏7S)ϋ)E !1~Ek@:?(g xav3ZNәFV̊ ^ C5ƺ 1ړ[O6劢q8ԤHԃ%$>[ͺC1+.#!Ja.8Z^6iGPJ l[a? 5^A^ުx9rG5.CY9=yJ{r|>;]x&sQXmT@P@ ٠Y 5T|IuN-<*OAb1Bkb U2xGQ|BS_"$( r7(ZBbJE*i)XZ* b eZX FRIP f !A7WXJKSϢ LnuR@ "i+)#7H{l ÅiLa]ۆܲ.uAXiwlGyjm#A9agۆm--OrܯdqICޓ08c:Vim"Pbm"%Jm{zɀ0R 夛M$ЎYM$05&L m'xX*Q 刮* AJ_wwB@^yTŬ!TNt<嫒 Rq6x+;8%-E6.cȃ-AbZ=&A$5En1E69E=u9EXoN7s%bCjDG/TsDGm7;Dh8y@ef9;-?#A?28(ҦQ>P*l5WIQI3_jE:UbLY3$"MBIBHþؙZmK# Ѡ8]DPs UaUmz&G(&@ F \pXNk7"]Ԓa&{3Dj|]ɅdI$#lr3WՎD-pRI >=Fw\բݻ 4ogE%sĿ [/pس\P.P$pDV8(`:!!ȳV1f/nUr2cZMz8wf81fH/-|(Ni-I;"͞ڀ65cOpi08:vd1"5 fk;Ы1bzoǤV0kZnKh7[2ҿ|͏/G 7yX76_`yȷr/6Y-STx$^bE݁,&̜c.3 sd,ˆ v?^bwCyv2@-~{@Ӓ4 zɧ |Tj_g,rvEuc)~Ⱥڷj}`?г7,X/y&Rh1˫u HFX#:VCo1l-cs$ݧݾ}+hWܿ˱Ç?|c7/e w3\g<5lbjXUڭ>?rƘJL4m7|\=UGN34aAMGV#9JW?16=uΰ {4X#u~trmyǜ1z4U]¿zS<>g9b㩋Gk}$SZI1y9xIo7>_QgaYCшv\n^n^s 2\{s0>/?lj`}8$򐀣_^yZ,wl=?F0Ő[LOlt5 eКh榱tA(kUfG7S+~?9^<5(Ә͡!΢,k=9xϨ QF"qњG慌i|KU/e"[x-O4rOU߅[rsu3@0qŨpH<( \JRu Pk9`;93+Ԧ\oejA Asɹ0(HC$2| 23<]} '!Wz[mvAeҿ_ҵKDϮhr|@8fP`'Q%"s :)b%j)gk"ň!vJ2,.PBbvI&ejz:|\a4\Ȥ,5}ljݲ{\EJLh 1ˬU?ќ)Yx]j̧[9Lnԣkof-:9oqtr'laϫUs9sN fqf ۼ*1{};x~X 51_qdt;,B&uswfT^*#g RKTIPsDZF[/w,~K7%+9ntjJ"ӤhWC#%'̑Α؍~+xc587GbGP9]8J@csQ/Wh_Y}ky!53r:{8x:}Дw'__d@$Ǐ+9`= |%60qDP2eaINgBa`VSp\,1ϕha Da˯Q1çOzM5&8x419]~pϨ|:QpI3.&z%:eԩSy%Rp1*&L L.5&8Rz'sMm3+LV'DU&ZqMTs`8 ~x+*1dPz8 Csn\1O?mMW铽/]:6>R-AZ);| /~HЃ|}?ͅYW۰ܜl ,ahHtq|0r0 8d!N;.>ǰ2fI>/f1J ʨ^MŦ%%S} >NS83m'|H8~kW()ʴ3h`Hx*EUo\čDŽj(e~ٻ6dmF껺=,l%Xl1rM`i1)|aW=1EqșHNt*7pgg&oe50xrTb6I;YefoQWd~bnH5p߄ҿ^`I_m'?a^\}],i+B靷hB!.m GdTh+rϧwg9// !i)X B & 1MR(pEdJ2FS?25ZΣſg+3bm;h?J_mFkm{oho)S++UBц7?n`9+8pC hGU3.w7iR_[g rnG6c@v$z>u5F5(MBUڠN~_Ə_:Xp2HSTw=5:jU5ǀN 4f} =uׂphVjƣ) h 9Ix@AZ.4IO˫ hy# s/~`^EaɉQ<>]lǣX܈P?"7Dx }ATU_o/g1fU"7QQ_gGXhs9!Zj\8|$zַ*thPy\8h:\KqT* %4KN:#2_~Z5Ӗ|='1'+.bAaR`$' Ee g6CUjnK"1pk5%RB-I"BZ"M`0`",4JJj\~!K) BS@&ц/-%[zq;^ ړ!0 V6L + 8I:r-"DѺMN(<|IՁ*4eZ@ks!y?c0.T'$޻ "Q*ReGf9y~'TEEH$[H&G(y/9*Ok%A\;D{SNXL)m_y? At/kI^O$l528|R\f㰂Vܧ wʨ2L"^S7@%Nm6QBB=L!J"1=e6KOv/s%,Y]N=x(ZS^SHI!qTIi󴋩O%#9vGe6 dm#[o J-~ Q4UTߙ"(/l4R]Xd.gP{ f9O%Rh$zM!^'XP- '<`lO]2GV\M@ZCD/rzHlP86Q4w:~hFwSjfaEٻ("N.5Iwy%ۓ֪=(#[rGQ7/xsJ91wJ9SIɏqz=^缿._.xBN(3G?+3z3?}c#[78Bey}bUK2 ż:e=$1*+}, ZP!:4IjV,[g9fJQ#)j:j]W Z)#u 5qp۬_<#WALѕ ^+ʃ x2QZ\ v߸^h0s?i?fB)RVl"#2saU*{\N+:O4|پ1|m#\#;,fiIS-< UTz?0cbeetk}or7ajk7n'@A?VA_@3u}_꜕׺mь0.䌈YB5A/GbuDNE&L/8 #=E;oqpOEWYGIщvt豬!ɢ{R~Mf ݕ?:z߮3}e^Z^h- QZ؁}(g)5sAצN|vicg5]H> pAIE)\$XbK [2'&ìA@iO~I%:ZPK3P$/Q0sEd.<{xnE>7HGQIb/|/&h ftA$BILCkԈ2^5i&/H68a$sE8K\B8UbK*Aԭf0ٺi YZjg29odPJ`\ Att""oW[ɵZ,< >C9Iykw;Xq?XQmmmmupX\gS|rI]L&%#7}?Ulz嬛"ryJ2ʂZ },57Zru(q܀}YP\PMM HH|\$*@\ >ES 5r1ZIjx6>Kӽ{txM@@" jPc, |\m$}6r Gɘjo(y꠵n4Ư‘FtȥI+4(ՆN >YMcaRw dgF ^F3єJQ$RED 5fHTnC+ɾc2z! nTKô LZ>vu,@Ɯv(m#KN \H$C<< j.L*E&8kh\.Q\餄LơwZC+b[vv5*i!iKwe'!T<=0(~W>{< TvGXqTAtzʪ.U/rS>At=a{K1ָ/&ٗz܃@иʝq@9bD$̋$&8JEi&FĉDOR-[F<ѨCd"SptZqSoi3hD`JE˂(,&)N9=:1Ku0#U E+a-̡Q Օ^&$%fy;Cb4N>ޜ|d[jY˟YD ~ B?Hۇy ˝Mz%ϯ4Lg7ui?ȪR`y'2Vξ =<}B[gNp*4AG#|V#+.3a${<`'HBQ *#*?~-:u >Bhш1ė.hr8FEHh)ok"=C'3!nMGT{G\JΥA4ϡ r޵q$E2Jrm=|{ȒV߯!9Cvg$XⰧWUа?˦ 9[RnCzv>B_i|'em:"t6Ph78v 0t}|徺oOȕ B a0Ty0f-j $*,ڙ&T/MUǤ'_]\ć&S>ILޫ^ l ,bu nW>cqÏ-)?X!M@$6,Erj"H+})?[е|rHNX4IfM2 T/[懒:g!KeBѕnΘ)kr 6?Ux\ ~D㣒?T߼q-y%-C˹zaf n-F[[E rAеj~ v3*bgoTLVCvA-0k; ބvi WgL'`#^8t4TYF(>̘YX8"wN!΄VHٿ<.BN`]iIј+(Br'2s\I\q<b^$1DexRטb\jG)=R h8;Sc;m?9c$ 6JX^-J ܃`xwRb?.Ec'`w:veD*xj֔ Qiў1e-5 EA#vhdžbAi !ZkMb2dTcH' ,N"TOL^6diNP\G=Zk$1Ӝ1؀0)MDUXK-`|#Gmx0UԌ'u,=|x93s硗;viQZ++Aڂt}*&4챙 2xA t{ ctƨUj]~e"fR>h[$f Xakٖbטb]԰CjuAT\KD)שvNL+M*JDRحpڑ8 9Tsʯ?l@$(!HTc_TpAJޮA`+D-岠 DR+*`-R-d LZ-iNaQJTXX68 i=aB/)ӈQ$gHNGAʽ䳂 [#U<0\wFCi˛מs]{+ǿ#*5(—O݁ #| nͰn>!y͸;/./w>2./qT^}y?Gxt*$dfOkŇunŒ ; EzH^װ6+ջp$xIT#ш0ǽS ; )Ws ]8uÅB s4 G\0B[ޑьp('Ƭ& MuQ:p"`Bi`-uE]л& UyF:z]5>l+Uӄ ;!]P.P$pDVFE\p"R J@ !k/pߙЧQ] v>o;u;޽Z639pGz>bLF?^[9s6m g]5?lV֟sVJ{ov+YtC֪=q{-Ru: 9 4bB@T^ e4x6Il԰@ޗA@9yUбZ!! k`ASD,",`i]jQ)$!Z:1==V]>oRf-;?-bgUϮjr* n;uT}CB4x&QxQEq" X$hdṖ"y`r3ZP8/uU| QCXYEQ51S:@-w9l,j.yrYimTTiW}jRB CY`a5Ѥ9"k>[w#c Fb*MtPIߵOSkM1*XSJKJe;ю|Mﮟɻ[/H7X#LMAK bk%`\M 7) HSD><'M^;337'U^b\nY42c887.P Hd.Ic;Ai`/@#t\fjN YKmrbiT8 n4VkB%*A uvD`^UbrMjP-z*:ҨI46J~qDEj@Y"-ȁ>$O3Ez#S =k6Y'H~T$wr'ދՇOy;ԬnSWo=tb-j{+((UndӗwvLebZ8TR'zGn4 QBM`bSprO=]Ls5?],om*YbG%6}A^_l꠭~QrD_]X~~^,gs~is(:.%B7O~oQI!ʟ;;ڠ)üC*TlR45挆=F#o3aX& (T2VvkGN'w t.FU9ZQ!$ac~ID8:sSW9ҽW#.K(X@%l* vBlww9(,:kqbٓu; 9d[nǔۙ5DjO t w;O`?1 NVpǯB5[?7?Md=.&cpJ"[oGhV?PhyC0,nΏy<-̼Ybfzvr6v1=/1HC&ݝr\t1*LTq6T`TN&8GRH ͼNWO{k#@g33ES3SG}V_wV)Al x{X k ry|I͌&nĔG*hI| GSN'Q: T#!GՉ~:՜ cRLgBN"8a! DžN'#5H aBFx Ң5j H1qi[`lSMW3bX)aBhz"3v hL 2{݄jQb#:(nsOLUO4W !!s=]`E0cv8xn'G6}&"P%:CV3>rD M]VwY< 1( ADBdJ0 \r%t*,PXaNpj\~{)<(Jr&1<{ծuzmbfr|ٿ}0ֿ1-{ lbhj6Oo@M+8YH-B9B"jN,!/޽*M}ләB/a3xvA)fv!zW VPF+p`%w,Ĥ::WLaŠ) Depv6kǷ3 (YdNAV"3-:e." i)sO2XXţW8b!K5Bly篇rF_"(o]bړ}-ɩ2ewV֭NPǻBUBiNŸ9OvI(L,]]o:zKj~L)Z[p^H,`pd>X_Tc "8XpEA,GuQB3?-WngN$AG:K@Jp5ev(JQ%ۄk>wnb$9&˭ԛRo@HϯBʇ/i'KA?»\WVzv{3⌲3yp ÍʘG_L߀g˄+Z=\g΂w31bFvJ 꼱ժ>cb"t]ӓۼn2ʞ.((^~|r}~M8dd΀Co#)ax^lv0Spv0%,L):}B;QqD3*&!1?aʘ`׼b<& S:e!/t=L6=C[Svυ2K2Q X,l$ Eu 2ŒO3XBZ"z%bE{3E~=.eo(w: MۥI-Aܭx?ZᮧcB;Xъ@ͨuhw #\^/똠D%{|+$Iup%mW iQ *)dow viЏOuW彯`#gWB-Uh(~6@/[833VZxtʘvꙜR^=3k\֪~zURyp43(ʹJc A޵5m,f!*O첝S!bBS4/m9 R%_t.(E=SDAp$qߣD&>HcM^tk#QƋ9ɉT}yetDwy{Sg7W/Yv_r}T%X|FvP$ gf|BXLJqչ_!3y g=R:WJ@(uU9 ]pŞE)~(Ɲs  )*' {F)`'esƄ>A^xvk@jԑAC}U]ϾMκّ9<Zs)ydgwal|َa͡Җ>OO!xm]TeϦ9V`6d >֭+\(#OiiY).:_g61.dEb+ˤOn}/6furbcz>/eJc؛CjƟy}-on㯷ӄ{eV$0|ƃc/ ײ%~l&vm0<5;_-O戰zgǴPF"_%Gx13wfGXa,F򬴩 BkmKƑ#6h>p!Vd\*ϔt(h%R1 JYXHE0[sj"4U)Bש@X %RjЃQs `}/fVl|x`ۣGJ3bhZpr6I" uz"ᠹdkkԿ I- {C55u_>,t^j_#ZϱUri& 􃬠WՏ/ 7ӗгK:)g $ (t<_,7Y5+/*NB/w&fdž=;ޮS?pl l`ۣ?`%, 5yݯqO:RRfc.!-8?&, Ģœn6VF/~6Ƈ7OuedfFǥ%}:uކ񣾨mo tlpH%8gT!U4kA=1\!04dzDSGʸXwgŶZSF2{"(F{Ɣ+e vhF7p*R$=Ixˠ5FKF K33x 3fl",̃g4_$EhQ.CR)l"&XIPO `ŕI3(Qa2GJR#g&v?%Hr Q:8]?t-H|CaFG>0--`Φ 7~@lpY (l>#Z 7m+1:0Oz?k[8[}HsNJʷm+vөV| t)h!~9XHg%U}fC]e48nPֶs(n_BV̜h]vZ@^N##?ly>"$A%Tzi@]"_t"0ŘKm^> # )F%Q lh.?;O2xnNJ%S8HR*))Rt']u&m3؈'Д<43OCI Nc|a}Vda@W!8EE*"6a.Z@]F./IFk|(p0o5EX-FznhŬb<ׁ G x)$1XpLce;Q)I J` _݁+]K `TKCӓٱQ^  Fs tݟR姫`+1+L q&@Ϳg/]79h@UwS?MX-y?dhoYVa? g rMg PZ3tgC;^1~gzQ8;LSK_C}@] ,t_Ukz85ȮWPa&h5 fRI|u8kWD7ݡK {UWw3p$SA=N|ws6z=[i)zXmB}N-זB'iAݙ9pѢNJ(H%% iDӾJ0$O,1RN阔e&`@=dT BXpja'Ͱ oaPcIvěHN*H1/諆s0R3AJ+PE.Xа˼}'Dxoj j[(. (F\Ri~_aJ4JPyD|h5{lpަɛuі&w-lm:3I0_&bKXMIsM@dt/le'cXTY5l"cC1ut2 v>)C_sc=jQx? 0,~)6 g*.ܔ"U"UC;߭Te=]ڷl<8yenyZ@]sEU#m],|m B .em'3DxRWJ">Ǽ~q^@ec^vފHiĄh0Ƕ잃#&@Dݾ^ iv`M40c.yƠ!6ԟib>4 i*;d)Zx.MSE@8s0}1D@5FҊ,H-CںDYWx3Z6|}2I,e W炣hrmQiz.%[O[ݻmQfyΛ.<-PNzs_0 V$Sڡ^n2p,<ڠH )SYcx)HONTcmNfZ_"*%ԔYm:FmqZ z͐dѨ98.;}mbUZvV,PCA?4,KAavxbF$CR.k ڻ з D| :L Ӽ1&q!&!C|}އ=rƤq{׶L ŜPA.xB# ]tbBj  %萢Da.8 &`oD4ET8RҧAj)2H@*cR"tQ tJ1uu#i,ӚRb5B3fTDXb5zb|jkL[)k\cml{ HɁEŋ`:;\֖v/Nr =lt;goqWd͙3^m}ٴ(.Ϋͳ<ת(j\Ț b5Z2&: uR'$E`@el0iDFA8f(5cZyϴRCn\JӂKĽ 3(SX+RaY++D N8 Rs}aՄĆN!bKb$f^] a!% 5TA6ѩѹp!#/snF6.n3_nzQ 7w|}k{|y Z/Geޏa\3 ~@ot<_,7ȿYu/*NB/w&fdž}:;pޮS?pl9*yj| f\qJ:{5aDn.[%Uмhٹ,8\ŴP\Uxykq>I8*<_xH2”&iJ\7d"Cc:x^u4}%e{d.KqδE~HNBc&*-'9&jA7V%u{m[bJѨWf#QԱD+fKL>['V}! 1tld QW.NRhxrZeĝ~(BɹEFi(|ՍFCfmR<#=j$,mٔ=}HzÑ#4!+@_~C;*㼂@x#?ͫ+C Zc.Wݠ<}^bGoG][?EQa04^cMh(~Ôɓ6$mkAz uCi^#FzH$Rtrɴ@?SR3e닷rn vs;&{,')$)1 R ).Nw"7nx"l5rZVCZ`y./cʛō5s'V˛+XWµ_/^LFK74"m /`p"`Xbbr4,ƞ䕑2%1׆YA Jc Sn+QLFapjFO[X5Z*'!g #p*ŀnA N@rUKn,8( .e&2 =yWyU& 23E +0iT&{VpC|Z&8rMEP5 dPK)~I 8iT#h[; +T2Uu"d)Y>lqa<;LR-4OR :3vasi7[6Ҏ/o6[1h}( [J3ޙ\q&.!4%w|*q"IǫR닝ōVV ˈ,y&Tw6K?7P>t_ß5j{P!l0ȹt5R 4.0GcM{3zf bNSQ<^cxCd(۞ҎV 7U$|NjR[jё=Rb*2UuմKj[j ْ9D$ev|=fg9pg3~>_;xEyx_/X>sr{97L *zٖ_4C13 6O{\CkV){IGN`:ɜ/fRjB a烬['ևٻ6,WfZ۩?,bؙa'M@b"Q %9)bSl/$m"ͮTkWk ӄ!,= B,w],Գʟ-U2O<yk2fE* !i9j;")bt"w $}yѨ2W;RHsUtqai;ęC}*Wha <S+X|UIieU]"T+G 0$9 !/ghAp1$SZ;m@U|I9v %ɴSJgMϦQh9BS}Rr!}uk-P0q})]^Ţrs8*J}{^)]? Ahw֬ͯRg1 #ZI^u_|W}YPKXR|teZphc!ġ o 29޾Hu ܏;y3'qP۝/S+6, ? <.՛Zro&\\95+3\=yO.?Ƴ_7(`/&F=֣ Y)'zS;T_){C/ i~XU I ij\){`u7^#*J{pK}~Ԕx6 Fד|q4 "PoΕ+{9 =O7 ̿xX x;N?yYfkqt1̥a.] sbX.]lM4ۏ'qr?2W {]xp'K+I@82ʡK'8 RZRtJdmT01O"DCH D3qo*:y0ہale;P!;{r Ӽt}ۏ3UKqz4VyfVЙlM_ʞ7kkef2jqaO8.J< aj UzqYӆ݁ C{B6Lؚ">6 ڝBՊd}_ޤ׌d{xƈ%8pvQycLz)bw V-L> hB/{:&v^3M#|P'YHB! S ּR1': !5A{L>9͡BM$x ZvjfPLg ' 6"r!ҢH47:"8K]nf.C/rjȩ"rjY4mF ̹͌9=pP$U=_[jh ac.EU08HOߺ8;~)3Y]k.Os\ߏ E%eqqV_Qņl3hG~ؑq+bëx5s}EwxnoGWT!.׆ltsML0K4LO]AIX9H 1leQvfb0zjj/@s{i&bP%BV'Zfd *$`-w<vIv,VLphhhB6mU vqR R;ey=eQPg1P;j=|`~3%2P#Kx =;/sfLRN%2}U[6+k5\g]>NjHe^8W1Km^ydazʝ[6 -UEB%ی%ӧ{Whd "h֖JYV<ڻ +y"$UOhH.9B,Q0 DA˛՜~WeUu^Xy[;)$Dl~-EZlH-+puoq. ]{׏kՌnخiMW'AyFuCjnEߐ6VvHLFC|w %3e8\ryd&&uMn,ttJ[Z)nx6- iWZ뎖gpϮr|sm>s3\L ̷WmIOvrv9rgr:ټ/^GQ]" :M!Ȏ`sjb_} ѥM{̖FAwv[RX{x-{!}S2(t8gPϭwHl`/xgY$DP'_]۝x_vA(ޏ~!UL1i#oFmyt> 暳1DYY+BqH.ҧ.\\,iy'+"q\(N2uKA9(I)p9cBG32).)?LکSu-(J& >, J8 .AcL:2f)uEJ}; F)gM5s ־ v a_7RXo@#s.Y&mX2`Td.v@ԬbfV>I\+K'fb-13~JBΧSI;5 B^B槄Hƌa,/P⢁WGzNegP貸 *来A׺I2|>d<8!Q8J 6oɢS\}92wt@߉'{S ,GMfXvB_1h J&2(݂sX,y0o >k!O7( .ʣgdJ.xCR^|#Jt+6g M11Qըhu=[#L;aӪ(1kviX-+,-v2++-_f&0,>XR^\RXiuZE]^d^"W(1!L m$DzQA`7fsYo&$fSZ&`RkNܗqL) :CMY&MHZ[+E8c() HBkŝN{AZn[4 Щ!4v(l(ȆiyA sc:7pIy4Dh" cWm ha#6HYDljua}T,N -9I.г‎d8/DQ uQ O#&POJbVJm4tS NJJlOQ+]s7WTr72ޏTÖںK*}-lz I9vILjbruhtFa]G^*tVVZ_T}D3'=g4bHe0鉋H⽅4ӊ 8Ke;NRpN${IDvfJqT;r!b8n;w~1qŸ8I8@!+?SJ 1E u>-,6{treMlZa*=FZ\H_HfE"Ԛ AY!E1RFPh4q }jE܁*~4WK(J\l컧u cHr y]ب=6&T)Y)|ה`X^΀Nljm&<%M`Yke&Ze+RQG5G*%y6R(/ֳ)zwu1FJI} +%_m2Swf]Q28~ V/'bswzu(I~xGRA_p&3I. '=m ̔}Y 3qټ&~Nd_#󐱃`zbUS!EesT-c<IY&nqN,t7?OA{-S_ɝ&Ng)۞M̠l]H5Oq=\H$Dt Nl]O*AeTk-zN)=x5S.h יdJ"ʥ~i?a"$d5UTBҀ| s|]xxǜ%JJqUF[Y[ՄjSFxَ[?g4>ЫS\t2'.G^KLw!V3ORB)l(xLU#SjM=6-N=s(NB!, =^0}Ѻ8-O+TPͅצ졐XM9L!$dB5*Kv)wIZZ2R_cv0{GVp Cda g(' ܒbBwKap0VƑAWeםr-]*L5z}lvnj OWuQwO1we6~fvmr%^ZT~ds zI@%nR JA'AN4@|A30ġiLeT#z#b+>C{9˂Qs+0axP)!A69f}>F# H@2g8@!X|ajB 5V H8yQo q)v! _kA" H1,%L`IŌX \&ZꅽGumkal9R5rQyaG 10hX$+SAgC:?-hX}L֜10EKiݻ~{NrW3uOglc4~Z4.otb@kf~Tg`M`O8LY]?6|q!Q#I$ /`8p(Өreʮ` QS"k ; 1f/yM6iA9C5)ϾDf1y a=Re, UZ]KScpʵEo#t?LK?-gazwsvgYXZMrr 㭹f]\1{`_b L !JX ;:@ϡ |ƃ;F%=X`bNO 9 [3e3e)86,QEYZ47`Z K~UӇ$.i8 <1-#Fk.nDžC3륟<&M\fϿNwKAϊc=;2OG]NXxOjc3ᦓ?^|{Soפ\fCq)38AaLt ur" O4ժt^htCq=4P({tO'kd5|e OM@HW^cT:VBkL>en='^)! N 9x+fAIThc!VmjI{K]' } lxim =%߮1߉E\}tL( {[ lWA&CRa卲\ ' WRQmeZS结Ĩ~1{u/(hlzo2I>$Nb1k X:vT4``2, &љʼnF*nnZ Qe?zCAK&`WqU1+k0K t8*:R`_t]KKUwlSOQ͐% mp+E#F[Aaz$2}^X l>݇/[\A a;YT1P$а:+kᄒ8+@&R!v5:-tXVH9E, IR۪eGW]J W]Z]W]'ȹ#tC'!Ђ>ѣMR@hID;Php9''l]їO4j,h/7S)_U.+mk_QDZ$wQѐis`\bYzi ƪer }4Q'5τieO1T9S:y&0N$@Q4S-D>Ͽt!-c(UDv;oOZ3HEݎGCB.gX:IwO5Sذ +]GI2 L˂T$LIz$aR1(0{x>ٽl]|? ;\PP)l]|{)((G%Stٺڳ)JF*aZ-R'44 fhN0!'h/֠Hջ^4$8_ w~)z4$8 _h0pcj*"4o!5UózҹbLry泙7ēMJoVcx5)3ŸSLR!kX̬J!Ly-!߶y,52(RI-"prdOU/sV"ar;}Ka0IfN;|+}Nh+ZHK8q`"$ R_(=D b 5Ezc[D("Ya>x}R ˧-IPm6Y<^lRjϬ6T0 `1C1TVA ۙNX@{噰and.cqR6Fi= +mGnx?6<6v4w!AqCEf5u>X8i:&d8ǥbb7lX2NL p<k-t=o;^&/'/.DCFmMoL=j1#FGZpNS2\<љ~.(l@ߩ_t=Q Rs$hw3L '?u?n;Zt?’Wi}bR4O@Q"CSحJōFʼn1s4[_I_TN bLzWm{{Px8ʧ0hckbx\l1u4GtLٲ[z' S)$͏=)j&f ރ1ckN@1vOn`ӝŃ ;,*t"^IfZLPA(oKN%HXNBJ02WvS/uBKcƯ{2/kr}=Ј u3g3frX$#E޻TKސ2RIC.*mzoBK6$s`xD_tvKع`nt8\8NrR骢Piz6LϴVDcװ&iĩ\~9{`aEF]sg!ӨM?3iiTR0{ڜ 308ΩQ=axbwAl֝ũip mۍzdm(w(( ĘބDQw,'.c'Fu* IMT#I׷!}xy_tTͮ.FWEn<<~5^PR^lcW}rBe+nMi6Tpw#V L e麄Ev}%,A*a 9P؃JἐF8^FPSDϡ PEs^8<5A5'ְX"ld˙Ւj,S( [c|JHM+zY8Lx8PUμ؟@~s]|wkTk'JS3Y<, lgK:ʓMy"O6j~7`ܩęgTkĥų: 63~byzX1 ~TsC;% j4Cކ+"d4E)Z4UBa]xE84w@jjI|;&JFT_}^h(JR/Era9reciz_.S/~%aB.+ϳ`\Tj5 4);-bY*kqkTU7P|99h1v/@(3*V. _7$!_t-톯CnN;hwStg-ftvkBBr͑) Ź!h\ RD'u[nk3>ⶭ˼'j&$+2Q7"n$N;ji\70ڝW\D)K+w^ْ_!OǭvL=R29ѸEOlX4$T=ORɓnEB<[$,҈yc4=XXƼ܇7! w`Z0eXGTMG[.9<1q8X5laiyt,҆tՏٝlCܐW@_kpFv,G*/^t7 +!sh!/(@2a}!j6[dڃu]іdꃝ:w?2ґ#ŧ.ۺnjU>FȨ4`_wP_s:[0"k7tEL]G)XkEc;N3}6g< F0T 472n9gcy}P7JYo}Y׻O*MF |^09_!n2Ч }Ч }W#}5ݰbsJse֜TN1c#RJX=GBt"v|(EG< NF~[ū|1ge7& 3&4Ϯ S%vԃh54LV4^Lq iPtq<0|*K OWU$ vrUU# s\!CxgjلދKe㥙 -YH)4Z~{~2x37. % eP+oν zt5>lGYFKկvxzZ{s8y4줌n6A D%+LaL<PZ. -FB]PeQALቊ m9qSE%pVh-)'w8KDMQv`O - 0(-ZNp}hFUTCzTפF!jJhi*DU(DއEi4|Гh98c --WIY|B[4X -vఆ xdl*d\N#qf-lP~z3IJZT+`]:c?\Tx6w%5eoB至`]\&40fp|=?Gq)% L?ϔSr.QhvGv~/#)P4Zm'BH V 5^@q\l|>9Bh#C5}+xwBM!B&CyQUc_ҖUxifusLnGQ`NM"ɝOdׯ^e ˉ}X넑)%OYnduw}:|k>5@mY4'نDl=ǖA U*a6I5(AqI7$5mb׃*Nݱk{i}AY0Z`w,Ԃ MѧUzP'w[Q؎uPl҉Ў Ea;͗k&ea\J֡l^ۂTHIux>fg O w A;!BˠAb{8\֠oc3[ԅw~nɗ B6X7_./7z:p~{{f-eh+! 5$Al/;L} +>oa4C;Lpp=2h]xEo}8|_}` trk ID%{y ŏ-D]2xEj⤼ex4QZsFշdpW&eQÀhbJDD FWgk(SCRvnǣ4t4!̰X6~?i/[,q6ߚ (4p[ ҐLx$ t-$)4!#0P%'*`/s R:NA=u$^2QQiHDPBN{[0!򄂦d|3&Sm4.T MDsWs; N0P`wC ͔Dt"AXyk-_PK: B7T5vL@H#gj)U s=Puj[0F5~V3Eԫn/ڒLjH*R*yJ_YF 5wu$4HiX⾸B[Kji+.)y/*Ƶ.)F@I0YruU:[T}qs_7G-P0")$%08(Qb%)ʂ:[QEQVZgh A1ӊN3F -4-|Rh A!V**YnMt(4+΁w`YRLa--TAYZ2 $I+{{~3D.C墓tniĵ ͕Իzb&P(/r+t56t`g[%Ix6v8aUM /)m-.TiƸzH?֋:ԩ6 Xzk4n'Y LrU) E g iZ&b-yj5)Sޢzb:B !u:弢(D `U5Ei9)jJ֢wLA9ER\$qST"vjJ4} lvgnv V߫`Cm( &Fn½gyBPt |.pVDᵚy-Z(*QZޠYy@BZmcP@dWZJ_ n%} M+SQ(_\|E!4ܬ&VOE] 43ZSߛµP;q<^}<=F6 <+n@'@O)ڰtHͷ)r.^ Sek.%WcG3Y30~z,-~S$ѸC^{$$Ҽ=o379O5a4jHyǵe@l2oX/4w֘R|~{w r-N>#COpU&9GzkS=Zގˣ':gë{Tʏ9 M!-nE{Frӝ3p7MQ-R-q`/>mcNX\Nb5Ȩ)IYQON)lj*0ƋIFp}mBEfUఆ*3-=0-"ud\+9(^x 3 G gGvJ=Hk NJ)Ҁ.B[SGhMҘ8$>ԩ4MԲ 2 6okۡQ Fn#nd,Svel\_<۞MpQٔ{MlZ*>EzС fUbЪlhJdaSi(Vc93& $,΢?~`;m5}ËHX̂ \Y5l8 ȃWLsr_w eNw7Muo|aPS!+À>w~Yd)Բ^,UV*6|%ճZElʪUH'=sOU0 yS4iNI־mehR1:hbFtB\6uKJn]̐w>E)m9i9L[V;73aIޖJdN]Ep6El TEwG9P1m8}3_uPjH茮/p76[2na\<̷d_',&VKu\G)c%©EVL\sn >鹑FGZEY@,{ۮFyL X5Ǭ5J'/5S&T͞R# ZgF)3ހ1S>N#օ4(=z%!DRqʱj+)z:&Q) 9:VmAе􃱉,_|)Hk](~_o8-„˾xO1A3U]bQc`?ڂ ʰ-#3j(U~}2pQs>L!" _5l|Fwws[zf%>#{U$|~)8hߕFmnU'rZh"J:؃ 9bʫdh|ϤZen N`yBs !<7!!7;g fC=,vX֑UA 2Gzt[@VK{%+>^Nd.nz1H]ˁ ?3cQrBlT9r$ozOi+D(A9 13.(L7MPeS FS`ǩQ0xt!kE[(c5EŃ5QJi'4u){B ?NM5h5M]=#nu /FW4u4*PHS hܯve쐦.^?ꠅk[NFoie9tdz ƭP HLgȥWpPD'K'`)0x@0@M`l{.|a蹪37f40k0M3Q& @%n(t1^6Ä8L8 D!F^AhKk$Pˣr,^r.*Q̹VZ>;7\LZ7 og96Ž0RX3SF=&#`.&?3s"3,Piܕei"p9dpGk4^IqcѪiKI8n\)fBB,4WPJƥ+%S+5Gvyj3[vMa均+6}Z8-1ׯ>%@Gnw̮*'E,b>U7Q*L)CQ/-f:st-[93qqո.Ky;t0/~}b}?ɬ"[Oݚ:O.T/r[xx2)*hfij+3-s)Kncqo ͋;E՚c4 wu^0hD"+9*+Jm[ iv3yI\B  &Hi .Rgi@y{o9UB:UatQ[N R]sfrBH%Ұ39gF~hq q[H5wҰU F(XW(cĤ*U*}|;Kf5j%Q'Aû:PP2R24?*gqS4a{+;@)lo|CA53׌6-l1'(2CW;lbUE{L߉'aDJ_(xPۿ(%5huT0ZKEG!3jw۝J2T={Qs:˯9v%H'KV(^l|4ZFQ"Fz1lj96?r)haå擡z5*<#.$S\/ hs0>NF5AmyANӝbDG 35DzOUc5&AO:P7WP>1vttJs sWHԵ AH9pt(@ϼ[iD6^oK̮ߪd86MRA9֕s,S;y[;T\ @tsj+YR*|nq gYa4X0_^EtѣjE,WfeI 9,گj7tT4- Fcu&0#+-T8!zUު W1tvŶkރU} ҚJf ]a2fC%+]ƄҚON3KΠZwLdQ5gi*ʓʵ.+OV:Wxu(3[Ъ̀+ԴuN?7O"Qlp5\f+C\8Жژ3Q(O$ Ϊ Ӱ&8|%>^C4 bg ;5ޝdjPa,y8sYعE?9 'W2\pt)0V=ϊb-VyrH2оQgKQgA^'Mӛ 6tԬ|L'[?RxJj UX0w*Լ.w_RV<4kw3+Ϯjix<-/os+|%i'sɝ Up';qUw'w|7OΫ>?}CP'LCyuٍJdߋY8Μ]߮ 6yEBnq%-b͎h ivHZSS_ڄ~]-d΂l_AY:nU\-.E%<>-|OFc_31f3w:r2.S>1|oo *M0Рٗ?Q|"@\ =7d0~).cՋCVKQݧ!-H9˿7: |ڍyn|.g64e.2_C#t{-kƋ%oN=)SqaVD ?{9Oi褆"d`~O](fjOSt1.t:"am?WM J\2~w XB4@k7?ZclcR;&"굉fPqPD'%"Wֲ\e^;g)֠''c!lVq" KeKIZFSH "ss ¢e<]F&>-\ĭ%`kGکi \4K4,[Ԓ.ank  h?f7y<(w=eqSy_$}b2I;ISDZ׼n]KU D H7wo[ñGcv=pM&iӢ.>L>iz3jvo3#J1{ƍLICCJ_]=/FGTM4JAd3expe FTڲI) br% ;N= sBqZ頞 z*pշ#Szb{a;!fdM%g>jQorN<ґ(Їr6轢x ޽bb`<0An;}{ܰ!:Pe(e:1x+<ڜ1N8ay}+YR0Fkfۑa&gɴG偅N:$ Dց^#Jk##I|Q3d Q3@&e$Bھq;4p{hlSpv I̗y k4!lyJ |/ UBn"JJdiKanU.S,kS 3/rn,ZO̍68*N 'q:x1~ZE9aUllfO^17 y˗ZQ80q)C.j抑zOd7IlH#(\7O&(W ׈.FtqN%CŃ4k(p`W~1j @ ڡgZtN}uR*T @s=^c/a3Lt7fPWpP ۲w/Z:M? i:-3qM9+RÐr&B |wYnq9̕ 6'/.uv~E8r~v:Jdչ2_J\*,x㘱2&+BUd} \F v\pԙ pl />ò$]fdn22r䥣:$lUrۚl+>./ϚMi $2n̑KeI3 c2me x R:f V0XZYZS(7SL5k!Ej8RKHYgZIvAN:btdnM^\I!uQ5TW/~h΍U}yȿ 4Fyo}L,?Mܕ 5D}Pۈ>z\Y/V+N+!_@\ On0@NiY_K < *晕0(9mg枭᫽{+gɳTOA@^,T`D~cch9c* +92P]S+"*MO> Z =Zkbhǯ{6t_S6}ۡw b$xNVJ1[2kŐ'V񲡑|Q3b^Rʜ?{WحF2ڀ L,r4xuZHrwS[ϪlY,Y_EOSdᬑ>kr`…X2c&qbjU&djB%w7]e5±gk~Hǽ ɘ;&D2TH NN-ج˷WTW]~^w r`9pN$ /0@ -We}iWDk)/ N1 /`SHS! 4 GWh\dAh_il/o~t E@ 8$Ũ DZYfEPql-9xk`S>s^NZnH hyR VHu0W4*fSX ƗsTxnUs,Hp&Ya! ) RZ&yfBRmZe `PeZ682AjV1J 3@@:y1x-4 AB2@jDES8Ge4ώn٦RF~R@(oYы{Ii Fx)*b**mݞJE(qq*n /Ew'LU9i:"As"2-ë~sm[u }9/\?+$ʩr0ouܨ~q7_MY$\Lvi;YX,..}Y ߬K~nGURvQ}>f3ƒY;6K78Tүj0q )gG4r6~LF[PZ[L7_mUs޳J0^yR֬ʻh>Us&AZsHr 145|H,L#6h9/9Ӧ#b9}RL(<]0JP=;gJ;ڽ3 O}'i4g"}$Z*sid QisB oӴ٫z+A;q[CG5y=~^AіTĬCΛ^IBpuR.Wa`mߡQShT1=/ĵ?u7_(z#y'>/Z*>_ViוSIɗ]ؓrd',>ly=sPhfqcΚdLeg?o(dQ"uD$;B S.(ə $ g[lUN*^Dޔ.Pt8L̈́k``(߼Ӥ~s/73| `("PWEʹWJ dg8+m 4p- 65|1fL2,/ZI$Hq;t,pP(iq$w:X4B@RکŁJ@iOvlFҚߥca/?E7X>kc(!UW#B+Mh)}d0}{YQ$@RMc}BpkDž4VN'DUKc5%w=T3*=DtD;i7z\8'QN/켑L+JE5QYpPHl ^KIW@-T{vޅ,Eawf΁Eo ՆJKI]?CZ^IvE;O͚U]kst,%\se&5m8 ؼ?Đs/.ߢOa;bzZ#鱔^龳pݹ7@l!zx"8Itn>ZY;hhMZ:!C[NRjG-򷡳`B ^*+DY#b~zuKhK``z,fJ[OnK[4_[''*^E=|߻^y% )yU>O!U>6WbT#[tB s/ G֕-z4llR@Ɛ^ K/%:֫Nw 4!GR;E7.y@8:2N{&KsB̡i8x,-}uU+t|~qH-V(_'1+Q# C\>ԚL3Cşq5(Z8Hb2WlRbVu9JuT7%ǘCjUJ 9 nJg:=LRW&tEwޛls돟UˆD1l 1<"^.5yǃ"v{iTr%;UE,dj948K'vd^ =N>z~VW߱:>N:L-׉5>=9~M$T3i*f1✴ꛗ:ȯ>Du,X;`"-+C4eL+ɼ|eDm\|N;BD ~ 0_M1b8kfW`ARc7؛?t8})3S@ ̳fJe?d 0P.t/ _>d&z,8sTb&]}diqy ᄄxHP="ֵ9aTu1f`y〖Lz*ցD&-2ZZ.*<U割b,ٙ&ʲ蹓@袴jT"]:&.)$8S+s˳Ue"]eUl:85'e5q#d.PR, %Tѡ5WVXEqN5!3*Z A!CTr46(!5 s Pc]u|"pNby5taVXk=YY P^Wt|vRv^?hd2Uw$eyΗ- }Y\nNuV;/GJJGXOaipm0AM_'SeJ&͐Z0+AEK]ևinL@=x?rhհ[VHd)gBF}>Q5?Zߌ9j-:=F@=L3hܩ/KKPE A: %zԐI W_v2qh߸*" lA&İVgfV{Ο~0% "#a4].֡'M~o!\ۖFb}"3M ? \qXf??>>> ~GcMឬx7)Hj*O-bҩ+H~z߿:IO=gfRډ\mX)?) 7G>,,";O}z/FvF)я.Ѹb ?ջ>'}Jl1}4æ_n,EE2^$Ͽ6{{7Pd*/OTZRQD`wN95*Ps1o`%M.TuƜ3qM 3)WtbϕGVP۠ ?I6;D?Ab+IgA3@Lp'$Jws~C[im~zmN} /.INzgZ䶑Ҙ/=(Evv N6_D+=vGguS-J3=bXG$tH|w,Ŝnl`r5#]_Z:t݇5o|ߓߞ2wvB=0n=y-OέG E܊q{v2vqE9zYβw1iq9_slz^mhqҙj-9Px^>͠澖qX/?-VX:VJ`$ΧIsléV>'MB+:-`}AY9i&ȡ@'[!߼銦4JZvvsW-A<\/s3OV#%VW"}ȳ}i/!Q^Xx"zOSC2gҷRX,fbC%ZG6F+4!ptVVh]=ȗﲅI`Җ:8ȜBRmQRV-VB0wt8O[ }׊`h7"ҮRɱm>T3N>^F %f6.Q7w‡t,2SsMԝ.·;T* ]SYN\ bez@7cBPE 4KPd2T  3Ck_Ғ>,|7̓nb+#\Aޏ{ō/^c`doCT]s%oi3S=1F4U}.Q w9KԵ\'`r)QAMٞ?=D\Xaӡ˘L?WȎ}H.CUȨVjrI 2?Ay @]s2BM?F/ ejԡh[I>):blD`?AS´R|HxKI5:dj¨#E! `Jp)KIf$/́ |#Z RinS'*JRc$TX"/Й1,-Dvh O - `)X^@楍jgAZ1U0Cp!Ԧlv,rJ$O ܠҹ0ZQDZmErSyK I ^VSbY &r4☬<⨬s,Yh9Y" m!t&2gCa412ʋ,#p?* ~rÍ ZP_oa,l2sX4a~1⨬y@C !ĸWt*^1VFa+ ^i5 _ #Ȕ _hɊ}{>*~~ުʟ ӓnq~X㻇xESeo<~ NA|4dq|};KYӺjjD)~5  HT[I2/!ɳ|~ngqS`IǤd,"DŽc$K*>lǪH2z Aˆ>^n x|S{5~6F?mrxWqۏ>ӢZs1I~n.&fD6rjfK_.H6 Dnc*HVvfT!$$$ It!E" "#]8Ns֬"`S8.Mr_>\~QMf__v!-PZej.:ַAחqҌV'v87*~![c?݀n'cWgGT7mQG1 2˒iҤځM"\H@˙* )3 -~^/٬Yg[@ذجaw_ &vFRBc[& N\@uVco8E ':t1ɒI4L9\lЬ8BZ!k䎎|oXi ƑG:vQ972e^^Br"j:v9Z뻫E1-K6~)U)N i m: $x՝BYi jSVP1EhH!<ՒgT+Vٔ1 9Yqg5QqJi^Ê;IrEuۯ[F覆?<5ʒ<}Lgn,"Yyp~5}Y+qA*'vL=،dG1W]4N"J?n؆.iB$E{$ 5v nLj  e7Y bBJa)JsMbfU^,Y 1cNp0 @bL7$:U1^cq' ^`F2vb<P}çpCj8uVwJ'_ȗ&|D -52L5b_ p${Nx.9ád&=ۭV QfrDA9Ty |J)HcK6{M( J'z1ΙndBUR:J䂯Vܙ3y'͸t.B-bq1߁TJI& UIDJ c%VƮâ,<|>OAܨtx C[f¡L! hˌJ* /rp dz=jЕQqsR/<dfA9.?4hbACk~s 4̰F.l=r&஬A}ߎ6:ܷr `Ew 4FFsDf{Rף_"wUcO6 EzKcƾ"BGP<-Բs؈'~Rmsj;0m/⋂9A@"I}GVnBGVnA5H;t\!5J;`#pˆ =;$hؐc R NQb&O r{WѪ(ᐱwPRB8/gĔ9N=BLJj ?H!8i\ڹJNbGf/'m]ыGXTR}>^hC|VZ*B9e)0g%e<\A|ʎb **ڃoy֥B "扝2yr!W$%NT~*j`ΟWdx%nkѿpJuqF'^&_<8|ؗdR ?~:8Ijgs[8N?{ƭ /9]I6g튓lp%/I9''!% ɡ )2._w@-uKj]ZEֲ?\ehp(TDtʍTI:+ZgxU?S4 vt6xW'зߠAzCj^9N? &s;˱JE\o>[JR%[[|Wójÿ/oA{pWٰ+x?wI%pD) N؆`hm.yPJY۴m(TTT?7 Ͳ^t4/v+^2ɪB}J8O^o-yViY%]ZLjVQFSRR%E9e& ~^KY'җ"+* yۑCU%nls(=^X c։\p3C2՞OhW;E!gĘa=MVT29ݩ(<hir$/wUxS R!@CmwRj\鞤$F0֊"FHML1"Q;c '`Fl7' \Osi? (#A:ғ ˧:]-]- -׆Ey'yJZ3 @=#3GzpŻ];+е5, ]Z`_Y:淋?1|QIw +mAd:s jWסTZj.rTd߮;qFWi x5 u u Ѧtzo ~j~Rtư`emۊ `ɹI?ʃȳtd%f,4%%3JIhfVҢBm ̳ CN_îKa 5 >jSJ J4;Z8}Sd,|cL@;@pljT {,BpZ5Ξ_wCRz$QF8olʾ ez3}h}:f&jvqgdb5NM*mX= qLfFC?e3G򔈦t0qI$GL-,찝7\3]xሑڃ Ǩ$bxZ@m!`SdhR4&ZƂ>\ m5sFv nf2GUml" eP_*LcOWFCy|J1m|JIǿtmVh2QM&ZAxcW׿]5DD|G3 xn-$8A|! Yw>C!l4'+m1ZunKzl3x J WD ]Q98䵶yG'Y:\嗾Ժ-ytIC>1iP)@aC>uL8'RMaսiIWD )A6o<hĎw)Z$pd` xZ2wf gE0#p`6Kі>n=͒6w!4[@Bp {f}4.4diȂN~cOZQmt%kOOĖޗ?tҗ/GqW@\Rm}#mŲah}_%4v>ZLR(¯-Z0r4BD `G'ӴT&!!~DODv Rm=Ab!6mK{m^k2K5Cfr楷I-Xˉ@;cB ŠF0<2.ijH8sDbZI;q. Vt{&#NuB(H F Ջ?/.S6M_}Ds)G)dL˫!]nP##jZ{=={q>oGЌԶd Ɋ J9 t/g &{#W<-dN Em5$3 n# 4YXEc~ d %'Ïz (J9kOZGDQ90jAcxv9WQrow^wKb\wKI}[!#mklBPc(e?0h5,Ksb,BxKG 9dyvӭiY- =-@CIXD{Ԛ.-2^bLF_i!9аwq3g]tܢ游n}>Gw4]_Yn<\Y'4f#'+Jυ^W+"Jgi!=i$y=I.dJ_6g71n]1(}ݞ6xjjWk=ui g>$ N2뼠<| `VT3)ׄ5GMdkr:l3rL@ghSnDǤDO3M "A9yvވZ'MA+9~, \Ysr5+{ٔ7ɊUj!f/䧿~N]H e!td4w#r5_D;M"QAczS3Sy8W o}] ?qmJX#\hXͤc]KU&lדY SJ,ҙ|veFV=!lD> >҈-FЊծ AѺ$GFG!h2Hg}y0g[blu<bG~snb nϪ7^^Fn> >T?gr# ^砸*m {׎+lQ| qйYJ=pErJ %|w&(nǰ"2"E~cJ9PG#4lٝ႑:>!ƝNHBu ^u~޶JSϏ6vsYML5!gU9OiV5*K|'_^QZeh!Tg[^'0//2S^!a 4믔v5g{A %&7xAH֘At]e:;6MDUE-^:A%ہxA,N<6TKq* S%F Ft*B!oGv:UMR"{ţ \q\#/缠ag^H 5ҋmW~6Pƍf,^O,?" 1f AZeMUٛ!FgO~qu+猩bfF=379b5Cx]CkӵKm7Ol5AfYv3Pn{q.5n=s>19C[&gd瓳iAڞДڳ;Gd 9G̸ifYȔ2n9 fu벃'j$L M;Nr ԳSm/m|'=&heTDz2h=K=h^Jt{-N0Χt^,VO0m;%bsєP`2qOL& (Cx;aVk%* %A |8@G.SVX%|u& *(Op0k Ֆ8hq[1bܪvycT ^*/O->=|qPTwD=ȒJ[8xT g}(0QouUݔX~ܧIlUp^"Lǵo<+Ш*2!8= :\N“¤U RɡSLpj 11`(&|54ԩ6IפIeˋq*?2U}teH3*9-C%T=rѠ('J !#2 +xsWu&tu "hcgE%ub^ R28*!{m(9hj`s. }h/zq@E{Qؔ>+O \3+&QPQ$&ي:h mM- x]kwv M'K|_z81a&lyt!ij ߒI.uÏh3Y?|s Wc+4cwDVS< Q"LދIڻs{?l)%#`o96BD$8@K'M1@skX.8)YkIyB"רۯ|_) Vh7aRo*ʡ>dPV@JQ8ɘQjZQwWI:E˙K$jո=D$e-~ 5Oq}K=Yvu[韖[aޯ4QTxCP=Cp1.ܟU:^sŻq.CFwsv22ȷc|;wGldd{c;Ƙ3䜨C 秗T` &O#YO#9MnG%BLދ,P]G,q(/VIo ~i|2& *XwVxk"!ayQnJw.g'_mD0ta0A9ixO3WfS0I3cҮɬ&g'`OJ'15x(>ۨ pdjr8H3 0c h=p.O.O&~"w}a7rW֕7jQϒZ_/nQ}yw_w q:oАcKsjcl&_rj0;whw]_۸lo_qQo;{k^?5uy4Zz֢T_~ V $!_Ա#'KM lf{m:jlؘrL;`ذѲ  x$w_K#$JXi+Zg됯Jبd@δs~:tu_"GR0Bj$^PN4 :f%JIF`OJ_Hyb'-( ҂A(Ct1@DBS>FH1Z(sNGk, HhD9L"-EHDY֐ h>"%!F+\m0A %D㽰N@ë-JߖlUلjd(U&Q&.Q) *'λ s'LPAa̬Rnwfrة (ٺ(6tFvNN_Sjfo b9+C7N4B1QhoqiӆxE(B}Jh -ȽYD զM,(  t<0=qjQ;qtki#{k>~ EḎ߄:@a^ QJB흅p3\5^GY`hH3CGCzlo,;3rD.i3V=z)#6,ՇMns]ͫQ_h~bFFshADq FHϠSG;t57 kح\v]>G9D͙>͙c͙zqKkAŭMɻQa!W9c39*6lVУY t q1)/tZtTt@1x(AO-W޵1:nqq{766=R]\^_ЗuUVf'J6{ڃ2m'io@i.n8ҘAXV@,Hѕv`?d)"dǟnNJo?jE ۭ԰-!f.toV*;p}\^}O ~f \!JH!&![(t l긞Q#!tLH~ +Y[k5p+b:[X!WC+.tߨ?H;4?v1LP8oMq-<ڲh5)p3 { դJ,h$g^F>3J:Ar>0YDL7W\}_-qk/:gՖp8!.nMԘQHޥyw< 1]/Ca/W9 )&`KijxP~qu}螅+猩bF_Y[6 qw i8!f1}tRgLkwaYxj6r-pBsL3w)ĩ>Uy=' {ȉ^(hy goa7`3pg.+U= bq4,lEEqFvH˪Eynr"_|20Ǫ{wDioe;h SYߢ:!Fˁ(-LFغfݸE䕰bbA͜ uW#@VMo^6{=Cw)5_׬'_; ˆrcP@gy3wy $1/6b!V+p_ּmPm/;'1׊ J2eSF0t8$|3t{rSB H~NX'+.`_ gVm S UԚkg4۲4B]Iu5(~^aMքO]q2V:VSIbHC*E h@(1L =j;p62kэfd:_|)Rљ NrZJΓ;h)XQCAvAYB1¸Z!>*P\6+чB!cH(oR'1OiipI$uɤ`N5?be*ї 0+-V+Is7o82iJ>ypqL CۯS9,vt>Lwt#DZ=j}]ڭճR뫑z'v]뫑unuP=~6ɲcw]Q4(Ft5l1:h-)qiB!u!M 3BՃnz[_8!?W:H/?G&}F .3Ѡ^\FL3Qռn !i$/C90מ;jhu-hhqK"\-:x#rw[wW4.7oRe|z|ЪZ.{dY.soYE 4,nQch4n % mw hTP1M]Ɂ(i$FZNjyr9q9A)TFP*\Ȝߤ rJ͎ @?~# +9F4åVNff<I>\5w oo~}]|~겺8@KetcJ5:1pD$|N۝EW+ԨT]آ1Y:҇G5Yߢqi-9C鄐/K]3OޢbneQ#wdî{uWhG3T@&ѥi8&hMJj=h^"/[6ɛG2Wnә2??*o/'e +ֲs=p9ׄuze!z?SFh6+e:%]^%9t}1HswiXX3J8ҏgo9C2MOWAKbM?&T|اEOi=ˡlG4%kU.[:!m3+c=ݰn^Ѻ g}Gvu#i-oֺ֭Vc+6Pf ֭-9C;X#L fZw4׺u1B!zؾKi h 탔J$z#AKg .zAbV"8I]ϓiy~u:xf8?O`rhn/p*)∪UDU} a^XnM6dUgu4S% l{Uk fqk9Ћ*A X ](gL‚2qD::q"O n' 5{'X#S @q9iC.';&6Z.Nזˈfi`e2Hl:P} \d2z{ m(3t|Ps'}szN'Ro&[,FуUZ9?[z;F $L 2\uA̮lwв8:R4(. TiRAu8u^AjTVT]?Tcki #N=%:/5\"J()]Kς'xiKbXrTdRNX/,.'z#[E\cc^b) +YblBEYz^KiŎw<ͬvT THb9vY:=fQ 7S{G@vۍ{ cK P%(Ie=Dyͯum[_3` !=k {>dkKd1'GSez-}i --󎦞 y®`7ZQ2Lʉ3BTz<8R >74Q{rASqE3SQ1>;\Z!H$""RW6:˹B"Lx+<-36qYQٶľ'I7|g}0w],R8Z3UP 4,ūk8n{(W8@>2J NS@iknU*s5R^{8p n$J2eS(]_ a Dֆ&_{;;x3ʌWM,QH8atq-N2^OKe2 ۋŒ=b1:=dʨ9yH72EJM|v \:6MV9 WݞW8|%y~ j>'Ҏw=mev"h&mSA׋t2@J}$_}8zY5ã 2|·k .yF:wuX]@\W w9Oj<&h2PSdJi JURiݦ*7і7?aFPaƔ` @YPJg%H Di%ٻ&m$WT.W _\[Jfk $A[H;ޔF$R\k84dLtNA0]GDI!B;&3cYJ$|Ĥz+SzjKyP;r2nhe9 7o}ڄ8ҍ8# "/Tqjby/ A=-\(lѩM?-bJI+n3+Avx yz%+B N@uW?UC S: i,*E]H_pg|]yӘcẰ!pe q fп3 Q!PI#J(4 P7p."g>_!A>%s 9FJ =J|2{k~ݻ&j : TaDr)4%O$8\hD%"Y&j6}x@"a+!q,{Z_%jᜡ dɁbX9} (pej#Ä*zSt_0!x10`VlR Ry6 TjDRrq4$7#RJ@ QΩlY%dt9j+rU!JTZI% ҔTH(Rr/H"@J<5kb^|ӹ@r2H ̡49`h W ڈfPj̱:cHR7*:\Jc ΫnYYn*Rg-ϟ@KkLfy!ϙff(YdfCLj~byќ\MliB@Pd#ƺ.Qc9dUs@O+z~{ /CK¬f\^n^gL*v{3D=m5h/F[6SO$AOB^C.x<\"0nQ8 Nf `V?;]IyqCAG%6tZhgܢ(p{O~nӷIyEo^ 78~UGd왶6tOmh ԩ3BJE=u5=5|LIEGjbTYӭ+\cܠҬF՘ۏGZ JNv/ve`QH= &lh4J[h>W 8,ӁҶ-9aeWW"DWy|Y= -%41 !`   Ap)@)3Bs$47O &D)"&<$H  os`FGR' -#_MlsY$|`о@E[aImƊHbbsJ Ps Vc- PH6n! * .)#[>NÊ6y}=rk ( >bNU<\yZπr42ON#!&]$EJIG)tQqqHη6Rc)o/۟Kmw>w #Bj/`;xu~Bq ƁH{@ִ _*<@Muv mX SvަgMOs/om)u@j*iBzFY|>}3>Dj2=҂4Hv \>>3F r 2vXZ>m/ϦzH1Z,{s*`tHvR=aϧ ͋b7x|COA/U Y37 rhu8nf}nNhu[wӿѩ˹ۻ3Bxz&,䙛hM94D[0^B jXQǕDl=n:{wP6x;_fa8찮5\#U?-=o^oΌjX>LVd 3:WE%k{^B/ DIII%ArºTPJYt'r NS_C0'5uPЦ% =R R*#L~FFLI>Xsdz^t\+atÎm+VMoMɦ[n9NB "Q‘ Ht3zN;Rm4>{$|Vǎ>4ít7w;dHʉDq$/\|^|JGF3|)setYhB VnԮS]g@Bm\Ld$^B+_pm -Z-**Kmg5z7ΗZrhA{)hcO>Ob)(0ڐ-%. x4_{_Fy{8 :=O{/5NC1ɶ mNBpwi6I^M3! y&`S8i}@3 k0 -e]#8N -͟?iN͚56[L:v~4MʐlX-hnWWBҗoqQbF>5Ƹi1wg{ȑu#*n>*;<?B*Kǝ K-4fmI4[#hжKLxJC#ҒhHxqB@E=H#0bP͆>%`ۿk5=<)_6,$~~4Y滙N lÔ5~(ָ=dmѤ7/~ey2H[ik+}ummB[2\^z H뒈 P`q>K86 <F|[Mv,umqLT(Bih<|u5 ^VROGuiߑ:-a"^u91y#kpk<Vv0dhYD%Znq5zG=Umq Wmnu7*8<Z"j#[  0h)bzPb~]xuZbkeg|qo =Muc`Bʇ|H4C!s5I,%y@ߍGݺ;;7ܭ۾ <!g3%>o<':)g!"Hߌ~7?~r?Icfœ,h}Rn^<1k6|y D@93@ZD*&30rL1G@*:OT  !&8%C3$H2ĽMH5r $0SI.d$GaXqXThD4)Cm j(ϒDX&453Dj1B3A$+SE-ӏ&3 032bœo{0svf~/j01bf ޏ|9{Xb*Mp7Xɳ.F*xq9{\Q/ n-^Ao4$>j5^~Ht6}bU`b?2<.liNa2; 턴ٻ7ndWym?,%Fl`IF3+ib;[Tk֥%U-{Ό(vbXB(y> "Dk; NV a pml杗Ņ{qst5?twF(D`gdG9z/nn%as+ j$v+m{ż53Q*nVZ)m둻,|q\iJۓ4Mj Z$rF+ΩȘM(e1]IL›?Z 뵿تq52Q@}LZ 0WMxqm=~:9Zo/rfn>M^z0 ^꣟@bpf0>x"1n=} G1\p@݌>Iq<2) SEq,XD{wRD{"K"8Tm5(+XIɅ>5+QZuC `*wi[o@j*džƞŖ "ZQ5LyKX*\DXp {IryjKkNouqy5^ 5 b@+(1z:_~|G $\2R\D~Pu\O*Çp(>< ]r05CWj˾>)  >fF~ >JbѴyjr5W-PR]w%횿,}ÏՋZD3/i8u\ p4gb1#b(塿P cEDa6AXK&œ* {O2'>Y(xl,/Hc )FHZVbJ) ĎӉ/>N,?IGz$]wA%j=w 5AX6Ti3'NS83XHu)Ŕ Jvs%nDs=#WOrWF=q^_l9C%rvv7[)Fd+5 [ Z+?+DCT/6XmWՈo_i?M/D`д$#WQHիIQ4AXz=xSN?$-F̦Ԑ ~I (f}xfj ,ZX!e#,7!|=3a%# R% <IaJ1 F82klAd 4ŲNr:Ln63 3||hH. !M~kfdr~\US913/>SyU>^a$yP#!mrs"e@TBK]Z:h?RRпPHeO% wLek8Ŗ\N(if:` ; WL ԅ?a;F>=8U!RB[&_v$$?@p~=PT`SCK*$ @Hp~3qRZ2؉.2 hKf<]vek1*{ b 5$-O E.;S/U\ :"R5BF½E>+|aD/C -b5#.벯S } H).I: o׫%!u:dF Юʡ_՝D1R8C[b S$2+K)bd vXGLjRtbޝ,߉BjPs~ojMWHuY!L>]g)ߖY !B?4NcP- 3R=ƜQ%(% K*L W1Z9nAKjS/ãbe*W+lI»>4U۵KŮ$Bo$ 8 LF"4(J"v t\jALm Q>2a,,2.$JTe :䰴8$8`-kcXMaMHia*+ 8!kjM]F4K1vT4C:W lT+T(z}[-UX5l" sj؅aWNPAj9Rs V p<2Z)t#хJ,ñ6KZ@`%)1)UNd>UbcEjuקZjY1TMzj(OޯLp*}ێ!}ZFAR&v^p6yfL0_򻠏F/,ln4[gl6 ;UHo) z&ww~Gpvi!1h e\rrx~ ^hnғmre% EL1[0i*,pصvkXֆ|"Z_t#g=ӟTҹ|dI7_*\ fT3Hg72Y +ռ[X?|#$2. f;)mc<-#?A>Wh%]8Ȝ@'MEH|J%,pchO_}D@凇{mΰl]#y7o>edGo'#uL`d ^;h#=_ʽap]I u|x)άs/$AǂR.=~;|>) {8ͳtȻ"nW[\xW[H1BXS}&Ho|w>Eٟr7͂S yx+|oW?pyz>wDڑK8:+'=x޸蹿Sphcg'pe2Lc^nTTI*y)'#CHs(NGB+ r2-q wlVBjRLh  v޸ьJx4#_1AB4#>x2B0W@6F"D:ڗp |,HEA~FJԘP)T0EL'c2R`K[dw*:U2f,q`X|7K2Z. ԄPBBPIIT2nP㴥34i7/GbL#YHHXzs &ZniXʋR^.Ĝd&qXIR,T~r.WRg! lgas0^_,CZ_N'q`Jɵ㯰h&@'x/iv!}hOm6STl_6 `Rtv q%h4c3(w3DL_SiO]\L D!_O}c~뛵8}586''eJwJβbRv9ƒP\О|L*>onFaqk= =mzE;ʘWݞ:]j´kkS4nNW3p֚+v[>2vҬWtd֥_pj/N#(F1T/Q҄(:ؑkh%DAZ)DBF*$HfRBAJLF&u,#)j"`WvLNE@ih"t Q#" 2TIRXiX"ZNa}vܮJޫ ϋn[P(/M :Kфp64nU9neF uܙ qR^2Sk45JtF a0Sb5|=)RU-`hpA ndP-8؄YI%J[2\rL2̀*\ 1Z<YgBP-2_QB)Sl'Ft)M*wH2՚,sNZ:H!IKکGZ 6A](X;g(ΝA#G}"2-_;"NjEfmHKD礥1h"bcR -8KcG KV P6 sD"obddMkꕐfL (J$#Q"Viޝéu_~Ndk>Ոn=p3 9c9]$ϣo/}p1' E~ Pz{p}_gUׯz~ĈK \i_$Z}h+GOQO3U *qN8*ܒaq8>yMSj. KR>eѐҎHڃCtC-:l+."x:E5}.փz=wTq#,uBcyj2 V!-O-όM7Sb]AuW4l:ajI#"T"ji|[jhi3ZS*J.e߾z!YȒS?53>_amviᏥy𛝈L7X\(YmMю|[ji݂GX!Dh3PW yTra&W&ɲ; z|I*}"C{;BbNB|^j8dKM܌!.rI5[E ?nlZ h\3VrIKB-!X BtQw{9VVSӰ.zvSG?:~+]C. dKͳNfK?{CNHޥxրXhA7Hg+3~4x~>>n _v;:a!XeA&\"Kv2oBHԫp?:ho |Qƥ l#.̢?@5p~hMAAQ )yƵC#T wB:vNEo:?._ j,;BiC=>J<힧gX $lNa Xát񐹯]0l_|2>rԪe}g@Xa2O.AJ}jj';$i/0ID_G[œ5$唻 S#)yƣhQ tb+ QP=hND|"Ju 8D@uu/jr2[n i\,6қ N4#pD끨q)Q{P=Bw^7)]7 ;V݀x $$]m};HOJB,baz}2zRrH:jJvdh;BC,T ~R }tY4)͌f3O*DJQz0S@54j-$ϖmDsqnU3wGΣ`:&9WH\Id&F[2p '0"ιwRP'U.pA3xM4v/\ 9xō8ޖyyY~G~V fbg&b0s<0鹍_\r|DMp$% ᘴ<8VkEoބth6-k\8 W;rmc3+X NI-Yn5\mW>${y*6NYD 7sF/8jxqh/->gHc3tQG1/.s }/^&J8F4|-Dé۲EQ.P{Δ ⟻jIy[/B2fx f,!5BZحf\L'og~9,{y1|(kA֮]Y.v>Es"+ #kULIǘeIA<;gXχ5]Ȁ.wژRz`e)MNAO\OgfRkQɾj%11.,ʢrdFV)=xpjsLΰ [3hskP!DzQ>(ItS~jT7EZ*(upY̘;byIb!!T[I$P$C2O"2.p +q'[!9tQAQ"``e dᤖr-.< JsY Ln8sY.LeI] jG5MT)X}[B\aCYz961 BE&BzVk鸸/]@Z yoAXA冸\6_,7ؿ_Mp^]x\![Y,Wp} 6*.GwWK;s|rNJjL9Y)/hͽevvXg( @ؐ y8D)0D 8"sުbĀHs&;3lH9b#ɵ197dboW=h\6oʎ^ Yi<`y9b˩QHJc,4,-ͽJhX0|Pi?^PCi+m]0P<ӯ]2HB)m]±`+:AikAo]TvJD<^B +yK41][`8V)ݏ"b'0_zbgdG?gbTu\ֽ .Z d|lLI&dv@SdȡXr΍~xruty[]TkQ{(+,/`b%3ŧHkWU$Dw{bC~.=KVso|Xꗣ/nˇW0$P`WȽ_ WN^FC L'`]-lޤ}F擙B7_%K ~.#-Z&Bpp/(ߌl2rj kh.>H,o5Ҟ|6 +8LkGB Cк}2rjW@fU>B$E}٬$3t%"̨Ho*ʸ!seݿET"FBHB4e-EFDm:? *oRĉfnc&#W\Q4fNqƐ!2a =kȣynr*<6,L. $Lq&dwiG4ܞ 0e6jI-OW((=xBPaD-Qjc-<KprUL^ekK@5Ñk0! H g1"1geY^ˌř@9xf!P)%ewM,~%YwverN&X'#x 7L~T.^.\Wzxwϧflg_'~q7bnQ*舻Qڂr/zCXPu u םyw΃QNWʊ!«i_=hMC :mOeӧe3 f5a}>;}d.@ӝ-iOWG̠zC)3'Z.$8| CrCjPMƬjZ;)6\ӚವڛT̀9cNbԱ1'z4gZ&2C@ha˰R.C` E$(*Yo#(w Bv v̤94ъR\,HaA!yhq.e(e) K/z񽿛}[>OLEVBKٵpe9%v[Sι@ơHY!fU.I0P`9W80#Q\wpw/֤"Ng7F1nP[h3hn f?_]lb6F:W̖셅l5up/J{3}6)XS:`t! di. f`\u֌]E1NIDҳeq Iȟ\Ddꩱ*Qb#:Ϩzmkj&$O. 2˾Y-7rZmM"!r-@ T6~GH۝涄w02^Mh7"x%U[ou3Ud7.;Wʞޏ~8%-y2+2y&^K12RP=eY89lUyn5s2̕3p\4A7'aH˧o+`;YW ~>4s2ZU'CCgPiBi D~p5apR[|3,[2 NR-Jsi_QEر|N++!~I[8e)uncx!7nqGnEc G,3!?m?{EC}2.roT,72MhӟM+¿?bهBߟl@CCh8{iU ۰P㙝SJYC1S?xYz_St X`lG^^@S+nXR@YVP>,ilA4 n<(bNsJK$7fڤBbU"tDժT?|~Go'+h_3tnӈ!M*<hv"}3ͮLa/U4ҕ__RX˅rΥS+_Ľ]t;THJFwdrv=JϖC99yrAǼPZ=P@(Ys-jz=_j_{ӛ54noE5;T0Am,(Z/Hf>(AriJt Fir W j-:d2>.I@TjrkF2 UIy DZea V5XYS2[LxB&( vSi !KJ)TjMH"$ B %Z4}`#4U'W5uM23?c ʪae &Sa&+A#`ְ bI.~GwMJUk>Ӽ!Ypa?%N ͩf bÉHb;}EeCHB-c@cUH{\ٚ#n z>pó{@Nz NbIԚ8[J%gIxK&Xޯ𗨑5mAc]=ء/ r1e 1~N+H!e|퇓GpO2QŹ:2-pJ%G Ԁ3 f&/lc'AX0L2ԇC[8rF1LCÚFQ1c5+Ц+`Pp,Ms!vTLKa>|!8&V`ȱDy3_0 3) t9B /y5P[Gfş+R6wB;<mC0L!#]Ecas J`l,T('&@BEaX5qE A<5c]^Z MW BC?e>-nʦֹ&5O;քn4!*|Vn}6n ^&-nnV"%g4:6B[]}BV*ixuXnYgru;Ʒz7{cFy̮G"P Ci1(:QZp!7 ޛ)]8{F4m :PK;27F$2 ɏz.ƣղXGX+,! Mf$utJB,c@@"F3`R>ߣ T $pUgΓ8ͨ52Va8&i X02Yf+h"r<9TkHF<'*ED3R=(E%$lc2}Q׷gdz6Bb"ZQh8S;ƭ^$"㔛]S29WfkX=p&`zv9Nͷhb01oy+ۋA7\0_A.]qlb_-Fn)~XwG~()9 ? tN!o7G|ot~;N8# >~`tף񧋁w쇥%Vޏ~G&,#pKL"^Ew|qC9B &Vedz+&ʹ66%w/b \mF|%8NmۨV9 AT% L2oqK0C`~06t|{p6xA4{Y)lW!^ҹϟ8`dq ɬ έ(]vVXN]% `%"?u fJevcN~sNDeǀ}h<[MlhEQՃa<61$97L`BôCvԃxjC78a>F?ud˽o=G\n[kr?i7ә-Wb#////^>!ncH88 iy amw,!1M KfTE0_Ah>]NH*iǥOjx@ 08E CmN[A#*fK'm8µxrC"z%ail)Lh5l9V$c(d2dNM &*T(R=isHLZ Bl_}e +^]oki)%)dsF^GJV2M7fDž'+1f+I܅.tQ TnIv]c lvytHs@bD9ؒ!מdž>M2NyLGXV$B@!EQO!N]#&JhwGn|`ݾR: 1~7=2FȐs+ _Zǁ58V3^讹BU L[/TWA*ievۡGxp[jr/b|w/m|n- r8J{ʗ_lX)S~81 %I bKg{]%1 2%ƦVў6]LWSGXGTŲҿgT. Xan֍œv0R?` 9=(#QDLcHPiO\]8] *9۪UpJI%JA)$4$$ `)X!&  %Bv6+,] <4I6ʠ,m48O&Em58YMRJa,?-*ty*JBms!QZ &jU)Ew'6=VZg RhC KX[:->D ݏ=v=}n YQW?G]ڃhZ,%}X;U *V ]_)qY{R ٞO{K}Tn}rZog qFco[բܕS DkٖzH籞6Y q`j'Èl "OAAZwSK=oH1ܬZ[RK|o7Vh <<ˋ~p54cwPq( "ߜOߘ7q|A>.ףּ{C [K&ݭޢn97R{}W^?&fE ]OniX-9ڵM|Ҳ:u y&Ǧ =ig:1h3k!WKn;Z; a!oD7lM XZa&m#Դb+;oˀe93.<2?#$f`K4GPћp$|`d$Tq)(񙳎V*Z a_8/Wtygglc^}[+de4\d ñ֧,sn!s\9k ,:RG%wZMp(Sc2Uzzdzkndӗ`sKrdmgK,hUdhxr4,-> neʪTR) -*i`84vl5]\ɏ-dL&ƅAq%SIْxY4:rAp/Ѷ; 9gv$D{GCW)!y.e}..~ffү4J =d7zydco%7d%Vĝ)$/-.sqsI_rk&EX~ <9$g20"YjY)8"e-OIX6`N [-äX6`9=L*eFHkFQ=|͔yZ Ն6×l˅՞ds̈́Vl'ŜI Y+d%)C=Ґ[lˆzw5ʙ4v㒖F:LR*zմBr QYn$ XZq׆6\`6ʲhiѠ)U ʧ!gV*AKen{P4RY|nιF߸17=vR`w[~%cӡ;u=ɘ^9 S;s@Rx`G,0EvL8Gʆ>w!%TP}tTZ>ʃjɇ|XjUUJ%%UYuV%a=9 L:>Ι\s݂XV(p+0z'hd̈A`/rvҕb\]ZFX0K֢@ at^FuȌ-­2Θ3A !- OzeGl3y Xg?:\e ^,/ҋ;-3^, jZ`1prr.o>j5@v{0 Әhww6y1b4ϛټs<1=>J ,`-KI(ex9*cpisN֬I0LƯؾf9֡AkfBʫČs6*c E 蹇&8&0ylN(u-n2&F)>)C~kUZ|^.Zr+Bm1#u 3"hTL)Z3)YK*PoErYR6;4R Ff:ĹF0 aR,TB\2Gsb^(\pdHfQ#;V6 sb8C;@gR*"X-F1o\ksZcm,i,w;[!jcVHcY9_oPXiwW4iITIn4>5HEagSao':8vStFC&| DiLf.Z6YaFoVj*ow@ǦZֆV(ak+Bp+؛zyԼ+9Я1,cr$$Q0Q.ENJ1%3T X{!{z]{k>:D.Jnc06`Yg@'W:27 ]FKoJ51\ :Ԩ7B( 1K\僎x.QTtCj[1eĔiv[#JD?V.RoޭRE}n`V£U]ycrˏ^a2B𹖒G_ܼd\9e) wx%fÕ v/>ٻmdWgφ#/ClmeT 4~ĹI !KRH3n4l2s4<̈gaqJcΔdhNeR_X e1l8UB))";CmJsL3iҞ:+rMA;?El*Ρ(y18.9Q@2 ނ%ya‘(#09owJA$Ti$=Z(wX[_6-:+H@xRZHd.:\4 qAo[ (=GrW@i<2e 寅H7D Dprh'Dnp#&(,"}b2Eȅ4+kB95SÌXAslsS\EfbX(0z,ܹaÉEU %#3c !ε/!XmΜȤ&ȵ ,}BQ8'p\)+ *4es.2dJrf3n\STZB0*M xB% P69zmr5r=*ܸ$S2&',2w)CS}{Ns" t&cN"Q,4p=TpQ3^J=}a&{ml]M[6 ƄhIQ]S7jCEWq"s+h v=0kud[ }'}[N 広ic4|e?^zDb(GADYMT)UPpShiJji8{P}QY /׽8p+&)-^ }6Z)t5mR(oaP.sqMŠWf{5m&rUJd>:@<5,pê=h}Ɓn]h ZA;;cE@ۓpu+"Fq0UEH%P tL&RA`D:,Gc愡*-B+@*7]"Sa*PD7D>jMȦL|/ع&wZ WWRJ\WjX  z .S(m)4>bjJJM@;C 9v*`68/S"މ C4!\EtJ޻MZXNU[wYn֭ y*S ʙ߱nۡuAĎQźuwl25<к5!\Et*1C DYbYCp#34TP&m _\I@}F1* @+0]J9mEcP& gFX 7bX.l2L٬ln lCTqj z5 _iQ c{F;÷zu)'! _r'#aX` (%UhVjhJ8upһ8VwXeWv⏝0|XzvS +jYE]|J>j^ #+&WX qH Ӻ\Q1bj!{Q1AScTAlsJ- Am@CFeߧwĄ DubǨb:@c[N֭ y*SA޵n\˃ Xx2;h)P{g.4!\EkU_0-,|- oWXJQ3DΘѰ)) N*wrrnp3$DjB5ϴHmEP-u:SV #̬3Ҥh((-riS1R%^%K:a,EVIrSbRqÝkO}x&l󉝫LϛNom37,`s?|^,$C3%:]Ud+`a' vP9sr%Xy+yXwtm1oW'7 ˲DsgMZ&It.xµXmsľsxA2'm, v!患UX뇢|fNr1I|3X峄h1,ƾt3z?*\O+aݚu? dpQkKO啻[ +.z=G JGS؝k13vQB+҅Pb&+rsb#NR} ")7Fm@$])\ PZĮXDLb؃bR:mpE:m&*z@lX/U_dTKFEP(0%&0ݟ"d!ܜB;#P 1O@Kr׮}oРЛr෠| I 3Iq_zq_zq_zq_Ž~&Ir.a9`kڊ\ .#QKdYI~> RY6t6Xcb[!Ych6eE^[Rv@qh@7Bl֞$놂yzb@>!BID=v,"J,ϐʅI0:'Y*etޮ-Td1|TBM?q\>o⌼Bd_gc]ؗZj-o o?)&4EJ0!fms5-"B*ٗ!gW'`fZ$oEFk|t׃aIBay2uR=#CDn`c V~7i Rr!Qj{|}ۻ; T>Z(_ 7!fѫ@{EHt+AfvF2 5QT^8>D ^Vakq{31u0Ӹ.=ph$QYiSZ],UJTb\!vƏ՘X:~Q}u1]?%o'ѳOc;p`*t'>P`-TUI $X0=Sl"d9N:a\<<8g902I& xl"vF9{\\L/"G& ]* ^)V7=׫Y|oD@S X0I¤n7& ,3j9#fSjHyy0Q:7iBo]7Sam4{[U4FftK 'Vboh  Ŧ,4 8ź{-[9>`e4o')Oh쭖/|53_?#?_yփr|"{Gܗ*_4׾m;ƀ a1Aۊ]Fm5Cl֐y:72YJsܤFTkBs"gV6%3|%4bf{4éآڗNU/Qj&gZoM5/-}Zm)#CTZEA&4ќ,$wdijsdCZ ՗gLT\j-*LKZUT!QhRô{'FZJqTsrryZX2I -m9-qhcR.´b_zTsB_Ǭ"0K1کE)%WlbD`5FKTLcXOl ױ]:*Bv\$F#&!(RpKnT%s]ejbm-/)&X !pwEBħsƊ,c"ɲ%rЏ$:5ic(O`57KI8!)rmTbN1ȨT8EB^dSK6JT}'C67p,Dk)X7oujɐTJ:0)[Fr efI% 2<޵u#bb=XE<"V%ٓoHQ8fuaHV9]TzxkveLRT} VQ9`9T b1zTb,Gبl;>"2~ 贷M M nlmN=FGp@a[ 7iF^YqQ%m c'?ariSNfQܵ7& Mw[ne):-Kij1d"Y1J»-!M?)^@|/$ArPDd63 ͧo@6]mqc6$/E7˘6Bd=2`߃Ǚ=D,b2uSõVNGim b\>^ Vk` ^MԤۅW~`6m37+1>ֈlICt )d JzVxY}kZz'.j(lK0{ ^c*唰w#T *P-VQe \,@%jP4{$ <HOZr(tv7b͟_| 7".N>MF;wvGg,s |']H:0%rQ$$6OqRK-V'ԣ9#Yj-\|Ÿ|ֶoszm+)V_]}>ߞڏsrz0^{tI~Rt1~w.ˍoL.|ݖ>[һ1SIoDzaJu_zqDK7|Wd|_*O\ǥӎ`v.0/,]oV9Ecq''㘃JOfΡi޴/Osil#ZK78:Ej^ZJJ9]!ze!+qlC߁ڛjYմrxT:э[sOn qޘ.Nޙm<]!ϒ率#6y,dw-oGG'ˣ~xX{`7GBސ0jLSetv>tZךRƂU9]QpqRYաRӜCjn"_mFF@~lW}d] ցMUP!BIW}=+l{OSHÖDG$9]xvYaW@nϱ3hUmn> џRRIO$hMֱ¶:xfHWPU{ ZB@bolLձ6;i#%m=={YFo&_3Qs]ZF%05PQCbtUk,HkfPy :U3#,E&?j#9~U$6lNhW}D?EttUkAF9\FG9#L %TWcTx.lkFl*0m=C$!!F1(D[F]`uΐQd-PgR5h?jY A`4&kcvVlrD@TxKp9 OvjfOԡ|"` bdZhjga$35/{jH9 2T;q`%ih/E0G+ʐФQ[#ӆP& KKIɖ(+YoX:dⰹ؅P*'uT(!+i. 1>cJJE$H!J0XS9.Hw3VhV3?)N?WV B5HaB~-|fT6ˀOL7Z_9f|3_ \LX $>vO 8!#]<2+#0[泀cl덮]!?E1;a_O$԰"CN]@3BtV|ga ]5snVCI I7 %T.Ϝ؍!gƒWIQvX'_crl30 於M5aZ.{N,+O^jMgʗuAG x1?B7!rӹ04DD[r6uuM:DuykW nMVH^# ;ϠuҞWy '1ϋ\*_ySyg-*ȕyy Nxr>)Y9w48 yn@ \GlWYI'1xj C6RQ:+Q7޻ϝs5 yݨo&Cjpĭ8t|Z M-{Zjl}Ah.7Q&\K]S4/rEpgȉ,G;y$؋9 ^zUtyB [EdZy#"US @,P!'˚ء|aȐ .lQ5Q30FP{}Ml)SLgIO J3I Q,mK)@m`OwbI2, ,Kk@/XE eƔSSE.Q žך3ncޤW-ɉvU@I9V,f?jAB-d^s-q(t EgrKi**P3ʆuIfuZUBƜY)U\Zx[9Sq)H2v@b?,ʊPX+EV@kh4-QV>rA;QE%G-q=/t +F( S,=`eICD9:2d`j?1Ydkzbv?gLR (:$m} 6DOT X/@966(JgyNcm$b#'f6rkb4L> [094qăPKa$ct[DX?XY>߉YVgV3Ӫ1itzw`McVQ̪phcgL-z,F.VSX;p߼͛]E(k" 5jokLIi)V-`]*xm&coj48Bo<2Wt @-l0;@,G]Oד :fC!JbVZr,}GB}fܼ&~rjNepK_C'Z?7_i*!^x-ޯM2+A{GSb+\|O%ir3 w߇mݏ'_> ˿Y|rDµe It/Uxq.QWNָUr΋rlBoȭtrxٻ6AXBȗ|=NpQ~'rɺ\}Q8;9GqFg%v;͋q]!.eܝ^_xU׫f ۛ8oPm_~gf*2 hE.kpc27I X.]|473PÀIZleVN'wǝ35=fM.WO͐VQv짫g ȫҪnBu'd{/_^N]f@ݔM?eڝ~Z=á6%O`]v}qJs܊nLEq8Uqy"nm3A(qx⽐ݠ sEj ,\Zsy~"/7YÜ%0Y]> Ye#P=|s0bʗlu1Cm9wEgl.g㷌0ypzB~Nmq}V3 }xL"oyrl4(=Կv.{k0Zv=]PIRE1^=>&ŦJf/q$/|]i\$^eB#vX_pCڜIY%b}ǼG) R eSJ!&1xnpyꊇ_ֻoͻ?oo}]QswHnJ7߼~T ]ơ?ݚ+epO.+d`= ?VOxsfr,;o_KDz j+P]lѲOЂFnx^oUX}$JGAXjZdy7(s!dݾMyJi_yYߕ[}nYCw@RKp1 Drb3/>]β똻Wv$z[ ܿ+7[YBw!m-6C;΂TǞt5mX{W|֚eRWxNDvG1ASG:hr=2nF! gF&JkKsWTdF{(5UOI*/rFT9YE%.FD16VO5ZU)eU!RҔ^WUr޸,$ }d^+2vdH1sOBU]NEH +E9* \0ڦЉa<նlEMхTAI(4@1$GR,vM=!BSDEXf6o%<}` L`h+7WjBwi$=߽׉ӆeCoZQְcEDP]H"&~`b y]ݺ\HOyJ0DQ 8y׻T]2/:wVRXlnBY4 >p{p_neyA|}nEQUfNsgܐnYmUU"{Ca"s(Wglى#xgooos ěn'!c{4ʱGj 77;l7POI<'=-x\0: gGku,38f肸79PQz%A2_mt훗vfj\Z0P~˸F,-70i 7J{SmkwQUul~WmaNA5l$ Ofe$R U`3fSl]v'm`8އ}N֛s JВ9n'$?ÄF9xIR)Jn:p {BBɊv4^| 9{w5J!FT}ߴAvmꞣoO{g(j!ϭeZHŧ/Q4Ru,I*ʞelV DXXŖAH |1YLhd ֽMew>-s,}imGPJrF aI2W\XPh#[6*,}l>Ƙ}MUE S->T˫`R<`KWJ"MOЋh=7E:ɕ1uc4Cֆx`繍rʝNq(PM0JB[x$h:^Im6em{4?nHsf#}*h~`:??i_R ]jllmT`,غmdW nu_Ͼ?efG6Q` eޏLh hD(rΗk] i tfڦ(9PBZLFE , F'Zy@tM3n%%5^Rrh7b{,yEzI@(-)U9Rшjo0k-n,7FԽN& [_0 Duʠ1X {>I>aL9:$ f%shu @Lֲ|b4HɂOݽP/D1;́') $VeJB1bچ8'Rq㩖DF.;D"wXlV!Կj(#YTSV«{d]Nޒ@ {7QDE%ԋ]PPd ]qZD/ ^da͍3R0Ci#g :9v qrdmJD$g5:>z#TS\ȤɢxBv=[B+f4PE}@A#gH$/|8c[C^pF]զh<[q/OlsP q<|qA0XUdNr`ZV !9=1յ ?ܙ"~YSZ{}rE!J- E yz5nsX㦇[2f((H0a< '9i!t4vsapE%0{Q6 Jm~}W)@MbW;EV _D6)x?E]Q?`o)2ZDxs3g֢ _&3U;щQUPh3D0nL ? "IE0il¼XQsQ|.]1"2`lLv/cB^d<xY M:sMi ;oӯ1~BS:uRLt [&90g%1C˅+7\ljY jP^ws:3eKP#x 9G-s_h("H.gL& *4d* JPD||Kw\pU@ޛXhR; `LhSp9~x" 7?Eɚ%!c% R* ) o jYg L7pgcVxOl+ƛ5i{#Dg+ƻjSZ0pae-=*A4wV9tc)-x?xxw>{mǍ#2KM]_ }q"HX#:8GHly*i6d5SGjuQPo/l}ohB `QgUAYIh*٢v@/dn\\X܅"VƧffzfYD1Mi" fϲ/% ݿ9`VSחݸ$Ab,i_~ DE@H2*X94AeFHґzD#1>>|C#)#1@j0$|A{*Mχ{):JdF,I __򏧲8o핰sr=#OW<v9G1=߂U!_Ɣ{N%la1FǜTD!LcMx!obe^dYeL\ytŽP@4ŗdjqUd:Ү:GJ\BTG͞KP7!Jϸ#y#\T)B({tȫe[3Ug/\'su2;mޗS\`68D,{+  xGC0( 8ZX}*ڗQ|kW>F%`,oT IZ[JYJfwE'su7\M['Dii5$Ij2_a4VQmPМ5:y-PdzuCE:YE倫 b͉$C=Rr588lH{-B>H<6!Xce,g%q$6/͚,?SLK.x*,"rJ+6ӗn1'ˡciDEGIG:!q{)BC|hDО܊CVO+t4'83=GCsJa|y~%qVZ";Gn>Jf=k(!I9c;FG3_wOxjtm~pzlIExSSWOƗ(ݗD#缎A |8hNyPJHR}A`Q=~RЋAcW,}g^HRUe8PLͶ. z j鹚tH(FD<Ԭ3e p樰<RvBc\zL ֚ @O H{{@Wz^Jή5Di[|BT @(RF%osްvD^L QqE08b+f$'/O>yuyrùy Zyr>G9D߇2LW@.w[z[HB5,Z[&%Ǫ59cSZ^)=V26Ldž@Mvq=N\[2bhR/bĝ7??;ۇŷ&F=&Y(w>m'qz˴jxmICJOd`#q1Jh+e9Ӎ?"/=fh7ΑQO•|H:?gG}Trq#}8@]\8ЋTɿXCT'C6$7VXq|vv>!Im , &U{rsōo{=]|VLG~c o竰=~z8CG«:t3_̾-w-PGcxx\|S>a~- tAvq^wwC x-Mn,SbUEWDh1R|Y\LMaQZ=eN.s=;wGO1W37GFuU}#+q ˿-H&ˆv& Y_3F'HF@Fi;eP]OS2ర;%p=K2%N'ݗynj}O^ rvvѹ2zh{pKAPQs~ag>02NBwb;ύ]ƷDP^>MnXp09wYLh(ok__"R_W.9C@n'ۈj7ޗsn]KpyExKY0_X)$!p-):{8W׵n2C[) rD} CHJ֭6ӺՐ?e#-놑8ح9S>YC[LVCB"ڑF4&i nqa@J FA"*Zौᬠ1r=j"8iiC&y ;)LHx$9 FQ! ǁ4D:L$LPJ∟C0T2KPKX{^kn`E!I6BhHL^<,Vs >hׅ:s5@`5ܢrduoje=]X׋hո~_k97wϾKϗ1qs™]uz)+xFy8pFok1ٰ{fN>?1:۞?]cl>I*g=xv=ft^fdRZUZ,°傽NzkUH0 J^^>>|h38FзSjDZ9db‹N2m@!ܧ233j肜8u:ޥbJɄ_8-_~ib|/.L 6o UwիMb㶴[!\-[Sjǵ3rt6ԍSijV 4? /D%Q E,'aIMHk]7|}螘y518P }"UME PHK 1;YGA5moGPڵTw0FH.ƯO ħ;!4 k]q#̈QK2>GeGUwFe8HFѩs^Hk`Fc/{ʢ&[-:X,[*HXL?_y7MI8p^'Ny£@FQUR:7X"aAk9V8^iɄW?e}Z2:MIe.-n735q4ᔡJzfJ SJn ϜV$ 6CaUw[%e`iV aČxV_ dݗ[eQc<VHО?$ie:i&`! 7z`B8 VǶ%$Wct􌐁!m)pw9%gH{ uTo;VO9#܎:pT`NVpOU|ftֵ?^( s@}Uqs辪>13諪5az_ըaPjs0*Zs@[4 A[+AP{}Ǖ:fw _JB;}p^aݪwXvJ ; |[GGb gRl%řT zNC'$o ]:t݊> %C*ۆQMhrc kv| /ue<3z0e;7- :U1NmsVŠPqEX=e JFɸ-չ'c|~d2}Ú!@ @p:3<n/">5Dq)fjyFZ0$ p@ Da /]Mj҉e N" H6R0Mc X ܸ@k>LO^4=N@И57w5V7&*e_? gzv|3 < Kc^S6^9B \WNrqfȅ1~q}LEȞb/7\$S*PԹRY]0f<)uRm=ƹT ="Q,G.lGQPDV70lR/1PU$R ȩgz_Xm~xr0t+#ˊ$O.~ɖlnT7)q2.~Y,UY(JSP@T@J bix.ecjqH ۿ\V-O/&|T+]W)*-l*oOOpO?|znv#ߡ~S{`U5qYT?yAा __~x5rl5)LF|fVf&O~$@2Nw=K<iH|Yk?7cS!H84ƐA IkeʿL~;o$+f3X<%ɎL9%-i!jrr$oni-E2*rXd,ʚK2f%ؤn4qH9 Jպ1nRyhlՒPUBVUqlT^kk =$_uؤ IL4­U&k# @J#cAT"7Uv65U(CU-"6.o/cU >ffY.@Gvіgh@&H$** h\j}{34l_=1Zmbd;ω4^Z\;ۦ-]!mJqZz7%tԄi$Eʾ eWp7ݡ] _jDx6Ab؞i^0>kM:B[՟&h ڬnKVylnla  E:$6KK7F4W+ 'AГR wg@!H:'$ c$$tFvGWL8=g}Wp$a1zC3Z{5z?au֊xIٵw:Ql9hX LqO5jl|2v1O{G1m;|N:E/9]t pK&J% v,sL!A.bCVDk7bV&s mx1zr90J185kӶ{u'$@iT&~ոk`.91Q.^az5,yo#9{[o8IiI}:Sxc|0 6E߿2r9J#s"6({>.DPˁxYɚ;@-m5ܜOn\ X]X^(Zb(鱤u%{ {qNa4E}eH( )Mj1"Wy1r[z! -z,@<@zi:yvv bn2!DfI ,4W82s~޾%;{KМuUYYf ~meٰTp[gndJ@'Ko KAs SL]g As_-aIs!I\;HsXʼpC9W^ cfK?8V9(X1Bp*J9dY%8fY9`XJLh[5M c.T`SȔ1@%KSQLKsAHQ(PJHx-/D[ܜk c6sm7&I}vASmUKH%<:Eb٦4`&f8PBm-C+`Ujhk;RڤJG֥><6r? l Pv1@>#~1(Cpklg[]p.T_VOo;(\+lD#&D[=XW lճ[=. x.]Q"}[ĉ]Jk9#`;[ΖO# a+fە 7 ^SW^SPn ȥx{N P/ӹw:^V߾_a`Ow{,ɊN~BNw$LqͻBY']DA!ZI}V/ϙ3^a"y,Pft~డtJ2zwFtMvFo8Im0aPϋI*Ǻ2è~` 㸔"6* ƃ\b>o88h-uNXNp'%2!i'N \4eT3 +T1C֕4`8w:aY,~Z$g ̋+!n{)*sfd]^:8G(xh|;jE~{e륤KCS^05V4>w2 I;m;-5m ]M;cEhw!3~ymxRRYqlu= %~]f^V}%#l$Pp$ kn.{pg^ֳ8.}sɡlZb_^xH>[bSjpARHF+5bx216wIԠ!>yHއ ȗWb"݃kJTsJD{&t_Z4uʑ .,ɰ[L 13Fз]4D<]Nj7?sp &h|7mFa)DK!y!H@IC"$u$c Ҳr;I%/~yqmS_ReJJ;R)f(#@;ZJ z ,rjL%w*A1`5`RH R2- 9KSRдd Y2<*\`lj04P@;/x'3DҶ❦GJݩsN&+=:d0SbB`[S_7' Œ42¥^a{ Sp1[&:,I,+W2ZUG5 qd9R \IQJ 2P%P C0fzgPfPC)1',waMt3e5&ukIjB']'?fwj~j4צjbRMiNܘtZ]i[iTSZݖj:{Zǧi}{}z0@SCK1/l}|fu?QPk3/zLO+SLRc߷]@lwwSǪ{_UW/tookSfWo %+K# aL%d4Re6yz-o\fMj\z6[paC@kT w`Z_Џ5TG v'!xNr*ha7jDbP1T~>8ﶂ40QMի烓&H1F#:ltpBŸ%4GO[Y=䞔&Dϟ'+dφWBg]qL:n䇂|R++[D,VŻƊN`1MG~Z೟qbD}7?~27Nů@B96Pm]ASkѝƢ%gpazD;q:$N}&8 @pHT6w+'3{u`态w5 t鈚Q''RHI`WZĺC{@D.A(pWxyyGts^ёUhO讣c#;c=[8!e2Fy#+u2pLb!d~*8Yst63t0!KjTCFf6a"SB jW2ӗ5Wmr^4ZB8ٮs"ǜ*Y'yͽb_]e z| =fVosiHWs!mĊi~'*sy㙫q,-~߅5?5kbPY1ٕ F,x_Yj:}sD] [ tf;,s@MT9~ywqR2tчR1?B@ r=HB|e( 3> =KAʔW'%04J0Vƿ9pk%2X" ,8a\ 0X l'ݩp5W`H/ýjJ˹@B F1ZHM[<瑱PclFǍoGҦYmzwGruSUuIn~-&Ayb4cns:S HQQE) AYy+=7J$FxdNT@\xh{k?@\9_n"S8 ~ SNٯ}J&=Cb?+dӃKxA@ =PRnGn|U)%lG1MjGC{"yMcQ~<SpN F1D#h!=p2B)x4Ter3 BM `a/ uD:A'vr?X,3$u̞'N'y)oF),iu+8u?ǟhUxub5j챞8NH%A#]L<#e`/$ N32ԩDH-/ɝ)4TK&ggx`_6P5LjDvW9&sy8_]*j'PZ[J;_\~Vå0-\IʩVDP]B. ǿxj,RR`0}dzz?^ _@=Q tr̃K;X{)08L:#DzX#V HClDR0,g]- Ye% eX1PP XeW@R%4#InKYE~}8X9n=RJY&3ם/e8h`7AD`Xr!QU! ZlC Բ`T^3FAFv 1MFAѪ".?,J8AfR4rX*- l,:IL7ǒS gY6^gQ9xwg~&|5&BvLG$ԭ.^2E͘ЇD([s iyAÅ8_.҅hËܳßDx*59;D94h #uKM+#LRӈ_ \:y)CMa #ưl@+p 6zM$Ey@N꠩`)d V[ #9 TB\9 M~KvPϜ* T=D){I'}NtR 1gbh3\b,Nj C_80^Z,,v9h'<81D ̀"< ð͔RD}\ _+]zZ(ƃ';}S& ÃBy|@D"&݆Bj@Bvb),`(NZ:ie0: `r%vb0(+Oiak;M`A.8 yg,RMfysX>8AhDwheڧ'6uhʛ:YCsUg{|lWWt.~zt-%D/P?JMEkYgUtK%"Yֵ6C>/,l"E_Oڒ|%'ӧxͻ}&RRgAmO2ߥ[(I2*\.|*q"Cx8ŰpYkyʐ˖*ڞ,䕛hM)2oϻ14 VAԶQ'ݪ;WnU6:};4Ѿw#-3zZ rLvj[~4V\ևrlS S*;Nha +R! "XC QŃ X2n2!):A"yc"ۇ? $o&g<3 k߭~t:-H>]._;̓WUܼ۟46Aeo7mW Y?eOS [5 ~=Ea%%_>N+ڎ:q܅}VR\*=SX::JwJZE=J6&ҺxqO'ڴDg'@(O'p:4y>&]q=aQ$ n,yc)cȵd4[d {OA]"8"h#~F4GAGژ229C!J^YmȡbU0Siz?U.\ Uë-\/b̵?&~`%.w.kv1my$&5AY$,WD1PB !g+|7߬4C:Ͱ:}VwWs? S!Ɍ9i;3p&)Qc޹%ᩲ0Ub{Qe<`0fYAcTiKc\nKA[L!/KJ{;H iu5Vy&B|c1xqA  D`%倸[eS O{y693jXsEJb/,nEU9)JM/m,jG,YtS$,8>y!5NpVȄj\8RR?k2Q^p5{FA:}p:u[0DGg}ITr̔{c3"H 8:Vz)vJi4G 3X)Bc;S85шDAalk590%MJ( /ӔspJ7Yp/ WH6Z3Lysi[9@kk|A̮ʷ~zsk>5nng=[o-F/U4*~KzRrK:vgg/a}R"t@7p$ȭ]U稚t<$JhDXs$X5: ,3309P6QY@R{j#D\pɵ/L8n+0|U7jPy b)=.:X[#,lCef`LxDcS&>qt|t>\c5/8<ýKw j gN%)B7nMf)X> ("&9DQ-r Tد( @[ц<[_ ~nBVS ; @5UR!ZY){[T7W_=,u)#yS%bFU<6Q}5D#{RipbCDQF WSh\)L>,䕛hMBޭөmĻq'tGn 2[Mʦ$L~Źb4VAԶQEͻWRn}X+7ʼ* ǙkIϒe=y YpP$`a/ uqjLTFq"w)lNe'{^rJ`,)E/+GUq$1 t/]kX2RרyS`EM= wt~>o uJtPUp0fphP劌* J)U#Z1(!mNb^sBAv '〟ZXuQ;9CƸb!mip@48EwZG1mŔu ' ('ZGnZp j~zG(E59j|j!r!O:-xMޘFê+nMR 9ݼik )Z/V%ScM1826ik5QrLjl/u"b6"5F+.ɳB] F ^ջTdGhE?+Y|}}]'tykI1z c_3 D ?|S&s mŲ~e? }ud>}gyHrk` j>etD5rnpP}Sf8Ko*#\Q YkC5O՛+4c=GwLgG7LTYЮh U)l]IۚE@(ۚ-Z'ީ^O.4oӎ 9D/Y(׽nIifl$M1(r+G>-!3[6ar#3B~W=dފy {-⫵%:܎=h{(VH 4(cz\W%39#^ Y[l6[R[~!6!zM[İlxfE1!:C/ϋu?l?Jopek[L\ɂ'|SIͱV8ΗUxo@szq1w)"8.]]vP,njJ8iz۳D&qJi5Kch*⺏$Rc,EXQ)Y*cUJlRRĚ$!0iUNi &q3& .4YJ0Vj_DG~YEa#YpݪFA\u0ٍ7^P +ȧncWM;ϳl/{w?Ѡ#Drjfˬ8DoBTj0O뚦'ǓEس`9S!.Zr2qCya"T`q`8SCɷ*L-{WamJgz{}a_Yv mt)wWG4HELi:C\G^ݙcU!s4X@,XB_\^xY!+K'f8&LCdȽw涵Y1L$RAj˹$k[̢$ǔmmIߘ}o֭9"fD4aQ3WaQT՛WCXu/IwhgA%r>AK( ##*$R3MQB;;6.sy0C^lEE@1+cT#U _ ؔ†9N]~ '!Uݱ doT\,32Pe$4sL"PŒ\Q -#Bqu!|$C<߾J nzx,4Ԓr-L@(1H Hk+7a+DX@QnűFEY|[T2;2 e8EPQ 0|*+~.ƾv[oyK<{#^ SڼAFXQKoBYeC[],EJ;b58\ 1x;N6K6Vjor 1A̞]YGYH~s=I3~r0ܧfz~5j}f>!M7r X쉳ׇK o ˿ߜ5׾ ki[pTcNb, Fm)w No8U8yRca$Q""ĥTIcY3  bBY#,poQ-V$Х@xu;!٧ڽz)}RJݐy%+K%Ig:=o!N^q5 k1wD+Ҙ82$!)HH>*j_Z\%UakSPzykKLu@JOS}O5E9Cҗ(I)v|RyfT3"{˖RäivBJ ҌjJyol)&2i7v| aRQ/{)}Rw|L( )e;~F5uRTX,K~4=Tݥr*wqNS{O\: ? Tn7y inlO@ Des6_a4%Z\޶tpA© rň6fVB sA; w%J.HDD os::9knCB^W&' "N RNsQ9=B4cS Q.Cp$|N3.ل/{i:-hCnSL4cd,8bq%2q, N56RHpK)R"-3# 3V ƮW$m$Ag$3Dw* ci91?9W+2Ei7"2;2SЉ#0|*1Tk9v[oyco\9ͫ9aqphNZr !ʐBB֞b)\9Hkqj?bDʳ2|"pS86tP{9 %<3-{|`HUi'8: MiI6c$BsCF8,@Cg"٧c\^,) 8wbf538jfc3L_|='20?79h jHThV4f|բ2W]U*9jiOE.+L{xvX|z:v0@;`q{L v^;Lkiá59L !JM#Z>p_iBwYVhe'؞N"mtn.́r}zb􉂶TN-'c(j'ZkqR - Nl(Y:s啙ӼoQhۋ(?$74#kYYG R"C+_0?l4jL΍2$gXoK'TftOu~?(7T;tI/W~pF+f`vhzϡ;j`dbܵ^9.A5eVΑ$oVp( XZl=d=b鵋T"DkWc18Jt6߭;_Nzba|nkq@5-Z'ߩ^ݚT {w*PzU[X߉o\Vvx @jER-Zg%c4tTIJwԋ$}C(Ho BU؛Gd*념v_oljyuxH> -؇ !kT=q5MR`&}(C' 襠}΍Ap{ww Aן#Nˉ]l :K\q&Ac@ϦUe:!V3:,KUM]qj?1% "ƈ9ĺX@NF1B͆8+]J65l : W 30Yh1szOD썶8֑2Di$ n!XDq7(rj4dmKQ!X0\*|+ ;^ \[̘5z9p[Gk"LJ2! k%XYiI (RHtfJJbg k㧉p2LeM E82r$bAlʬቖK HTTL\-_2[魯"31n3KEa㹝=$˕aA~~R $QqdqKy6Ļ>et{O3xH7+~@_] lXCa9&}ǝa?F8|Ӻ(N Lk\N&n,?CyHp "D^c{`P1Ă7Zeh.=?bZJnvz vz{@rUfz&9P9HE]ܧY['L̲; ~z 轶K-ķhwNb8͛x "F4ohzѼyf6zcj@'ZRokK;S zpiA$8Ai2ͧ 请Ut4IvʐW^QwZ,g;;{i|T~t f@CL!Jg)<Eڗw4u cLfgO_~S+nD& ,@/ۀXMbc*aʒ8v֣q*J&bL*e2eASg&00t6xuo%$@.F4C: ݿFe)5HEmnh e&EV#%1%kۼbNcF~4TL3F\`.d lWN}~'n$RPB/QJ({v  ƒ4sDxcWa+^F V*! Tvqbh]ˤ=MN[:QTT8 Z`SK,C$IgX;F$qTb"_3_NWv|,Upv{揣r~{v>Ǚ>8SSqa9Zje[(|ѫ4}dlǮ?yJtժT ;,u;[p{E^-ME-6g7ގ$E6gx%ΨF1x?~U ӛ]rA1r^' '#g7R(E_?Q!"?RL7`¡fLv̗jnC >M bKh#0x3v|Fi:Fv>pֵvk-ڐ\DCrs [S BDU\(B[<@ֆ"ZBDeZW|2Ռ ~ Xe¤2_'L[ݷ)n٘I\+F5$҉hR0b4L1(˒9VrC,0”a'c5u2꺶.xm \o +S_W)h),LJ1Rz2Ռ`JEK)0)t+S)CW)h)(LJ)|>L5 _4{R[?D>) Ts W)`) ѥs,0椥$y7azqWNW(0Q(j/tjc |P($G1zZYNf?ʉd/a h~<Zgg0,t}}g1zVRj] z8zVNFճqT朷^}RΎ4Odai"$ Zb.c4e3Ilʦk"!|OS(W4Fs-}v@P`օj(I+l:/P}j.#T8Ж ,mI $:FQLcԹ>f*4!o8TM ze bB`$FF5% F;  o)XkeOzS,9/oy%SJ[xz,L"Mc@5a8 Pu}#_{1iL _dd2w\ bi˵1I$ё1cX8N Kcԋ-2v mqQcNAl9P2%ta\)k+WUtĉvz}JwEe)j;u*pWtzݍ!121CqUO@r v8$QMύ#_'`H$ADq J `}'XhJs*(NTI 3iA(8piFo(@LPaj F߷`(LSm$<*ݏ+ ։3 PjXj0/OCY! T:$ SRJJ҂be-6SQtQDw,FNHA\L5ׂS~YF}I`-ϵ[4^GwjP/}h]ȄM(qw[8 "& }j q Y`L]pm0y%͚`$Sτz(w&(p FBrT!y`giGĘDdv^'́Òg#ɅO/ EE MH^( WKF ;=1ù]y;x[v0dl[s aI&#P&XwSh@8Iv#.tB[ns_ov}zINaɭua|[80ۯF; ґ0fYb9t> 4P(plRdn#4b!곛p؀1I,IdL(8e'S(aI\ ~%#a[Ь)hszTy9\ѫx# L4et%V(scLe 2hHJ"# Unƀʷ)V02-Ag1i3'*BTF=j8ڊóq"z3_|ёMQb7=:tf~_r~@|{l ka*cMM+ W|ekg$Q:R^{JrN[g4|G-^H͟˛5 Kհǟe=W[u ch}^|ܫ1:Cwv"x,'Z0Cs`Z=6]cQDƹ/μq?0Ġ1)F2.KDž~d!CNXw}:bM.tȸRvՍQ7}#} \GmrbO0üb\b\b\agb\AhW%B1 #`Ne{g1BN15]2 C聓,=%3)Lz/,9r=LL墇ms'zn$1痱 qU qP̓Ɖnx0JSOx~Gz;1sIQ 7z ]syiƖ=CKDbRav)+jTJW*k\} jWBI"$AÚK0ND t*E爸AZ^u*>:EIQ3ެKuRaݢ҆N5ڪSpm NqQ!.TUXEex% 6U4XX"U(9ĂؔY-!SS* ISAгj`*iƳk099|WAnSi"#z{o}rpyTYwe1ef/PN5/Ӱ`PHixb$hjf`F |ҙ3rNӈIv4.|ܸ<ADdfB.dN}U^8a_n]~G$=$$uUskƓ&+é?bbx\rNcpOkxiB$h!֌n狙؛у>rPdH!VAf K[),Ja(LɭB)Ir T"lsB(VV `œ f$?Gㆡ‚<'Sha0>`LRK@w+xwOWo{"yjyֺ܈ũFBC|$o q4mc8 Nv+g;|-f2;?) <3^]Oqь\UX ׁ 3ܿOe; n>4:{g |Ro )CM{1s>gC;A/,90 rdLr+5x*ۄfL/q˕ 8(:ݞiLgimGBdB] Â;"'~ɘ "B.3+$Z.rVPHO``k55ޢo/4Gc`k)Bc1탼2s +3[C%!vsPEN 3R+`dZp.spd&nv񯢥bּcj7Nnzw\qQJyng F,]=eZ樋c{̦QK(!Ev,;,hulSݼ8Lp1h-@,7~hoW刢 W`7>_%[kńyw)q2r²GTלEse&Ҫ)_d*Ss%ؑL3ʴ,r"xOڣ;\N H>)1%3cUHtR kvV-#p&\5@Qi:]*nWT{W^vm  ~؋v)u=)ZѦfQK֭K!&6T\ie!u޾p*H\Ŕڪ"hb_BgOg$HPsB'cD_6x:&E\1V-{:p&=*`pDiA7q><uy- =iS zKSy{ Aѻ [iԲEi-MnlgT mT}uZvrrZs )PәC6Ϙ)|f2.v1 K9X cdrN B2,ɝֈh :$2 3ܓS)"K0}1Yx+z.rĵR¨6 _H]%\2[8EDkCan٧ apݪZ>[f`HV j㒤^[r2\k-go;hJn4ǬEiwr7}Nb tnH}}ckm4&Z,LnUƼnif.( ԨXQ.?.LŒdA,Mij˺@IxxTΝz84FnO;fkJiu6JRg.lbbwnֵ zJY|4k?ѹ7Һ49~EƜsg ;@[RNP@^.D'@^^u*U2M-E;wR؀Lo0IgW1r T$0L ¡8 (l0L"ᙖb=@?Jq8V#tX$D0UtA3ϭc.?O7phK&`=e'T}E'aa0^{5V6ɌX/ҟtE,ofJuI%A<Q ̯|8-O~/j 7]~Gw݋ݿtr%ϳ:"&drO\"-f)4EBI(3hkrWrxeNξx<0w.ų ų ų ų n*O w*A T<3H=_lԷl_UK/ؒZ,YjU6-i~b4!`]:8X.;XeoYJr'i aH)) S*]n4b>ZW}P%_K" !1W9H5〉v $l |(MVR+ M VmBACVQ\PY6gyN *ǃ'D3jr*aR4"hZ }h0Ӗ;,9XZA1(B)P\Ǹ$UZ``u U8wJbNYjbE CAX.m* ]Eˢ@pR# !1DQh/5@+WMÄ319`es;STx]8q5s)0iEp V6-+iL~LQb|l8_)fqjn$=:ğ"W~0Y8oGnXAD5r;@Ƿ]0F`pt4"Lܹaߛvnx~ugښ_Qe7"Ta7l&d_&5GJlͦOefII*SLͯdI,wOqXGaƮ%?ý{hM1P;[#sx{>TA&AƆ9ОgF f,HJ1Ӫs+DEL /Mxaf*G~0!)a_$8>eL- ]7 `g_F^6,)wZ &skcJj s0e8qj&-A$J^BLv!~E!_Mk{wMe8/צHk5۝Qgd-xM3ՠjifWL>F/~};^b0Xgaߥ;NuzFLvZʓwsMbN? !bs|;6淪=Q1H”i5}rS1 y)CMyvf /eg"ΐn >fY͹ik9z=Eayԇc8`BfvmDžR9~hqdCp_o d&pe$xub]ڿw\>O]1f[z ;,k[n#Jkһݲw񒁯Pi)5`eWO?]eO~II勶!H}v\ڻvз !T­K-VT@۝}It AcwSLXD D'SwbZn}tf28B ad2N **E,ɼV -GB)cs!8|zpvnBR skutJ\eHP8 I!C F`̲("?N?m%HP\z@Rf`'RK'TۿFVp)*cJ}՚kn%ݔ T|ZkW*F+{aa& fᛵbBL! *rFդҴ2d2,8 |>ݳ;LJӤ4AQ4;U`8P l^R)Nq^'usđL^q.DuO]䫰N!-\'A>w;?T޼[`wkB&nSacS``wfugs0tEyIB' f(-c$*Cna!gY l%x\c<|X4/JjBE1~KnyY)rԟR6>^G`*Mnqw(tTu%Q8p0٠!X8'1Aw+ v֎ශvś`_WA}YϾVYhJEɉjԑ[Ȑ!4|!-LO;+J`||шۃ]]ȫ#eX0E&opK'bl(WzsLƘ`X(q8NLHuBDhR fǑQ ߗx^dsWo`:{P;N '3壯Dq]/}%!C/>  z|3?/NA|5Iߓ_w_570_6a/W7} x[>.7W e7ט%.I -a~Eq4vp!q\ 2l>ߤ,6˞TÄ]zOI5Oxz:,_%BƤc_> ?7şu>;q3q1lϝ]ΆDsfO~TSZUwru"%B\MLϢ_]WY"EhN=&s3tӰgyvvq=N(SIM9i@jD;L7&QiuY-<p}jdYhn lr*/Pk 02O~>}PwtH?k͝I2rƭcYf]γL$"5\4 "}~%I AoK!M?-Y&%?k0BG_GHHCt##.ϊ9JUF?C"q$#-@`P +21G<2^*3E UBrk1Y >SSZ?0JsӎX$ " +9WaqE{k|,.֪5\!,HMz'&5.D]Mn4ʥL'RL-p|3NK- ‡y"> 4II%\%=G .4숐W\=E ! &)He=rSr+6 {`qr0#T2,y+;c2+׺WxwWÉsvSQ/z)+5h暫,CGJGx.0x2BD ScuC\MG"B&ӻPr(౯ ~d2/Kx<\X;pOK}U1ع3߭=i";_Di3^QYAc?oyU{JJ#DF]}^Y]mPjj 9zS\;.:iN B`K}.5Avje䵖!R%xhF"O]5Ǧ)y k @u7ΆDrR /ΦwgYfH=_ Dv 2'Uq Znk{݌>U#e5-՝EK/`OP{*MNEI#"5V~1蟫*pbfy f 3nO?knPssA?x8kXwkސƠk*QAYM=,b!Y|RFg1ﭢ<' sOZhٳb-bY7Q-.W Wo|Ʌ|Z pD]j4Dyu寮[%hTJ|1g'ukqSD z0;sQFu”^c!aG^+*}ٹ/&Tۙ? ?m'օ"adn,5{Ɇm#oW*'OjICKN.h,5R0yvSj*12- ˭8=ؒ8+5 9]|]P9?N'd, Qͻd )7ogM2Ӻ]+ܿ.UΝyqDx06dϽA5Qiuf`0wSd} SdxF*Ny5k'%v%Z!@i<1'#!RdVDX'/Z6zk$%;D~ C ;8WsK(xXz+|eÊI) T.\ڐZ6,hK3XXu91|=@nGSsZP}t:^]ms6+rW2U)rJ*REf$~=VKK9%T⑥$`Mn֨6vs& ?bl؁Qf33B-9Le&;:lUD8PΧh _l{~a1K(&1?UJ?owLƏEUݛZoDVge  >MAFQ~8})sY@3`#qA$#7^ѬX=m27dFS!љNjhw?^3=Ʀp IsVt}?x,, M5P{S\?S΋rIͩ.~rH55ZmT=U#)h7z*BճLatcĝh(u4fqs?\`%t5C 2j%U\qKu4:GQ*M,#JQkgЅ4"{}(yo=ngܰz,CRH/RVP{їRܢܔՕ}E "0/EYBA'Km,lSjG+(,rV0v4 V{~#";|N' ~3}wz 5/ڰY>f׸u;iF,}"uv+նjkj[?9ΜmlN~zt51HGVE2ە@} [z[I uQ>0HyG O& BZO&Vmn$&I?huqu[ҿ=<9=첮]ÕRHdGm|AsRS Sd{!9鋬j`#̝yTliP9xSqOQlQ_^K87`y°^C}[ ktR̸@catJR̕RJ*%ɝ!['ae:WbQZK6\P YBI;Bt L EҢb.̪&.eM@4^Izk[^Kf=ٙ nhX+VcQG})%fÍjwflE``&hLfF0d@ff քZ!P7\@f 5-sa9jƭ5ttUhl*yږT"ubI*-&E)a״]jкYL5!eB=!43m\+ލC7m?f>K߽Oytߡ_lk&Mh:ZfS!v)їf1Xe#z8I 9Sz 4Ȣ]qV4A? ʊ2>>t*'uǏWX)yW2؞V?ٴ-e%F;ow?@L12KBBk7R~@#ݗEjvA(iu<[ڳsM'EybI(2 +J4yLQܻRrbfSxV~ RYDK //49*!HC( m_I%H_1=KCsWmL; .,<$20N{'UiށXrL/ves*cX-=i}WL̲姵k,k`wOzªy" Kyެǡeg&P߮n:*k?W}"^?*z {KҠ,sUMF1 Q_&dPLAgZVZE0P_bZ㛖bF+0-<ӪQ'_x L RmZxJBVjigZiZzX+LeвB\* /5`! J!(墉Έ;t0jbdZd A̸R[_S*߶jҧ{#L?Hs,HxET>0;iSw|``Y#ԁж"pƞ[IF朁;k6.jfWJ`srv 9jn Lb֚ÊC6A4& Ĝ 5jtn ɔِLLmPŴ7- <Ӓ2m(Jvs}u*1m&|hg_+xG2E7o>]?gj̛1T sJ|iK2ͻ2.h8=/%5j_:/|[_%H|26|KΉ̓ ZTF'ӲiX>>O{2LZC7g0O}Vdt A#΃ڦ-plsdj`혅\`;VwJVk0eɌ*K+MZGdDYdqb#Fqs]X.zzFӲ,:#'R~bE7tC$3C:T[?bLV2f>NQ$%&ܨAA|pitNXzKe BDcLۖ&,5$U[%N ȒQF҂a|ږV$,$Z4x%6LDř9x[d.y^Byfc'Sتoa/}̝+:i9E]fI&U2yl#dRZ Aޱd V͆WhuVP!,ixuw d}tF~T{c\<)_D2>*b:vPUD:J/)L+`Rd.ubi5X<"!'NͿ1/p7cRp#-*UY4!N'}pjd?'_z>sg4/][YHc=뱿Buq kq4[+VB7G\Y+;vBoGߒ9Ji>,~h)8jX)GZYA\p 2K+//̫s,}^M{gqeqL{}CWvL!bo1tHv0oo!tѢQE|r"Bdɯ\H0H~ c"vUڡH~҂Hpz -GP_V\3}jg vB4z7 9ZIC9hwo^~OoXAtv+dY0ɄEFaf&o8oa.J8r5j!ѳ@, tmYs`M9r5 w, i3Y?L+-YY?żPFBV)fvVun;GJIva_;+|W&A7 gY5^,JiIn7cQP6(EnT2<k5^A jCދ<',vPs>PMGd/!H/wF6u-4vB#`:)PFH:  u4 > Waj&XV?w~f&bZT2 EvD iٲ꫔S~69[][od7+F,0)J}qlA\gry& }uNUus/CEI{U 6*K|lʣ>=gjfqP8UPZ_=60An&s 56FZ,V9:=}GYu8sZwʈ6ZF \6 E % R/!޴si\/ Ħ-i,t62Ŗ%%& *uG:+,HquuV)p@S#+LP|p,IbYJ8%(eWRm{:t:tKvơy:-rLy&3J\%1\\88g|Θ1LXZpE M a1yiqlZq;Oϣ+n D>Hbb&@oԗ{R=_YgJ<;X\M$m?&;+.Gu>.:0;EyASl>Tٯ֫WQ6o6Vw͗K% ?}7,m^}Ƨ~צ,^4o*ZݮG?Lo0ʂ?}ҁ76(o.>&믷Kyrߪo18Gf]_wU;>QJ;-dgGȒ V_j>;WyiZj,rWju΁O;ցRg|ڨ8iڝ (fc*$Uo}b)Z?^ݮWֈu]%_ۻ͔[>߭?߭{bۋE.[i2lF|?vqP#bڜq&H)p`D34w7,4 \eґs4a[}ڞ~3%#\=fǼji@( V7uUo#gN.~YFW_u>،zpG ٥#(xn> Zk {ҎJ.jz8fZhT2QMB QX'Pa=cP~ ~YKk/7OnbeQAibeՈޏUJ0j|N9-BkGgs]Əh"wޘ놺o: ũIzPߕ q ʜ.ӍT ɟyT-`:o{8h ;۹XoGnrK@5BX.1XZ^vnh/X,`ҹ*t'=uM꫊ KEA}UYUYϕ] (/n3c nܞs}]]xzy~Sst6D-J?TgPNU[Ul6!__/=ٽX7L)C^v)-_[WNoD .oSѢ[nևrS>܏n8J0uAtFLtrAVѢ[ևrmS"2)Nː7䴄 ag*bUsSR_IG֚-ӬfQḱULgX@v%t' gqqKb +FZyDDw◪w}M^L+C7IEmj<ԛu2(moiܙOe.F3Qz@3ԣA/pA0ت(?v }U[t|Wa8lX?'7g'7iUW2͍2[uZ,X4*lz(mV~?##MM>*e-ύvF@<wO - '".J Vna2)hIőo|xNހ<:ix$l F >ۦmY<<UI(=b.0 ZcEiJC|2b"I W&enˊ7W$/U)iƤ0I$W1VJ# K~tȀ9Hs-T_tByP$ժ 292C%c!XHsclGɣ\c<-d 35=z-Y;";J{U6|(B ⓺*gPkmּEwۚph F^Zp,jI;w:Yb9P9\3œp.{NDq5NV5+c.}XowSങr>NupzR^dmZ#.KԖʬDKڃ,(HtӕY!o8X><[ grWjq4g"OfdVx~MntN}y7D8, Xi'/Xa%E,w^J,s#AnWcƀbzZER\Km\auV1^D)5l< #/j4B!F !3>haNmoaAg L'.}?͒#dgdNm*6:>[hcWHsT94@gCutjvLQƅprذjHqdLmY;S7OnbԙaQpfn#xV3~7> S)꽽z4hP46ңöMXScUX#_vwT1/aiEpVݱaE4Q:JlR;Ŷ-+hRkqfy8ʨsy]{M0 ZvڶeSq^J΁oBȜ~#ol¨a<6C X]O{э΂+h;~# {쫪;2a!E6.!G7a= ؎:"(v=b;wOr!Ew|[(+6)D7A$ԁjDBtg+V,Em*ɔOI}+5QJ쓮Ħ*`KrWj 7zGfڪF#Zg.2,U[b"T زI ̵q&s>DtX*Sٕ:0+EԦ^-559Q$H5yfQ[o&T9j1гnOK}+5i}ZO<@/5K@hjk\ia/wEj4ު bGr}!웄SgQY3Q YiTfdFq~X 1'@f(4>yňX;po$R*YM>q^P–‹$lTőoScw1;^b,S`2q6*sbEHpkfzCXf|A^g׀v5pbxs)͍.a(kuMFqӌ68Yzrޤ@yR[ :x[\ KԚUs!sD^mL'6KEU$f|dیt\qgmE = Q3&OUjȖί&+ iȌBYֹN"$@tU#^Z6/DPĉTux:,\@RU$?par R=5BaY3S?ԢT`01ZnX$?ہ^l,2jS2i sea4\iT : )D#壋6gh`=L@1W\[M-?3J00}FoVQiݸh0r0},Xн%*`%線YJiLzFy'! *lpvʊ9 Zj6`sWAvvt8ᭃ1B񑛬NN).HR_E']tlԈ,x#؊+T\n@vDʃ'[lc(:MSxjS'P{H (1U_mtQ-25lYb BB8AJYzx$ʃ(k2MV)ڻ|0ѭxX^(,-~8@|2-GH1<#~Zr8Aq0f׸aEm Xھo#uh_ewߥ':t-:V5nNȆ(ckMG{ e'>trrR-6%R#YvA&ioN>T6ݩ=6{ۇrSZW&ݺ u~#&MxZ;ARևrS3ν覵̂B\7b۔$z:u n}x+w lA*)Ž<yA?(| mgQ +s):pZmh)4,# ڠE K:ހn/o.BSS5=+YNY$|* :*ęRa׏fƹ{XaL2:r~̫2~" ֽg!ofg=Q U\%_%wEvZ﮼c蘁-?;! DJkB=嶪7Rt}ffcCPٱFUF¾Ilz*Fv&0 %I|[RO'/ЃĘoNay0gmݨtWM`zvwYz t2.l]균8i"9k!8.ju ^ Җšjs~mO\Tg`&Y^5F9',hD,3J ']νŇ|W?|iX0&!!{?[V[ \v#y y HSCUxOO@I͎ _P@8]녵Cz5b K &:Kdq$̞4#Icv+i A{)0R?d͑qfB"Z(6 yU 51 oLAD6x(Q]G6Es&e S_VbQ Q#&:ަ#}Q/d/O׎/k-]XV&W5\>;)5lx=d [+0ڂj<*&\O,g9]nmtk^mW$Ş E Y__ыjOxƗ7>0(Z٘Yo3lX8r_<~e~qYӷzYLL$6rg^5& $- '5 - `e:XQ k;+ ^)Ґy|^gԆKxΨ73fT&P. TKKpܰ/PHsBJ *,he\"ykql>";?@x׽ы~-~Ó6hF곇߬ft[=_}~q1we?ia(N!*ۋ*$do"|c<G %Yx`5W^sy'yQVr ? ߬_%?􁽻CT`b,dB`8."=Њv7F8 2M~ $nHFqH*'@VǹA>5e(fs+$f@2ͬM!89Ֆn,aNy=ϯi㝙nBL tQ %⒡V@G)$->Gߋx=Qlj)ci,70u$&a/ӈ ^% C z3~ D 0AF1mZ2A#!dlUy7`CT\ans,Fb_!,DE`,Gc_9DmV,>j'Ncګb-YϰPkPHݩ*#S-ȕĈ=6!A7]ԙjbHMސTc5b*utMdu&$-\uMeWBnWwWⅿugwuQcST}o|y g"}~^\W!s2%z֭OHűsσ/^CP?(ggiȟ\E{TWpsFT [_NwnLIf%Z!4OtJM?|nLZi#s!'?dE2':)z"nk8/7@Rԇ({oVu,Ej8 Jv]-^q_,b{De/{yƗ1t#(Kfamɏ]}]M!>nݛ/whʩRacY'1C)̔`\ J*Aw>yC7nPdތ!,K*K 7LI-Rn{NurXЃH ><"QS[/ $EPN%f|ڙƈBr vCOoSF2Ώ ԒPOBAl{*g*/c]ƅ ާx)D\Wq]awag\X5~ѷyVT} CywY"\lD$n%ګ#򥻹E,nU Oo6Dk 5|+">xϴ0eJ*!s*6 oAٞw<4ZAv] P˱BguTr/I_Y6dDN *<T6߭2v.fhd eaaL:$9 .% `_ Os[Hʙsf /`֒1V(b)/\cp'nZ&ެ-j^npc۲%BC*Izbz,7`H0"̜ |n°2'~_3B[p^*mfgS(ǬJf ӧB?KuYBʎڝ-M3di4C& NniDHf1KV̑åzegȥ1/KiV/{Y_/}r6“޳398aw*"mR^eG{}lkт`JZ: VKaC6p "y3 ]lTpWvy3V)I\{Uj4LΉ;<1c\XՒ0hE^<1= SzUjB4)ݔĐ**d4ӗCTT15ktDB4MWE}nl^hΜ0e.!v<]") dƃWMKK^ .aiQjRg롅fU59hmd]KnAAk;Y0]0/^@= wào d*$ 3 S\'Ӥ[2Szz/`"6CMb)%Ys#n,i9pC(#ww+2@|E̘nF?5DtlXxc ^Ɓ>DY%hjpL}jZr-~?4b1aDz[MI|鸩Q#1Vj/U / #y)X,R8S܊ Y _0.T!탌Q64PjA$K!(֗aSZyEKm[XґCPLavUjpλqvo (M 253 JӬjY,.4fջUOY\iVEu YF׬[*$4 YfUTùb|JhH$4ի&ackN]VwDgH}ɷtb؆ZbP쌖dЕpеn 'sɆb8/<I("}z\D+dDk !#k|Apy1:uΤODAc .1ҕgd 7kF)= ~#/2$#]pdY6dCrMZxxw&&/GoҜ}!ip*šAZITx .6 :^W Xf<*BŦNӧizc/;|imAppEl<}DRh5 }ܺ{TLfCƜ6Pyy}! ;CFҀIҬ647!f46ˇ^56[̚5fa 0&YHiUQk'ЬPՒ6P^$cT{-sUte4NbQjp\Ѡ&)!WM \ 2VJ|JBLӬYs g|͢DA'n1Y`Tf'=NјuIҴ͆jz.ٔ V][B4J+$~zjWMjIA[![Vk(XVSP-kX[!O0)j)Z$4O ٖhOQؚX4j} : ?|鵁֕` 13-S"%6l;^zXg"ڔu꿢]5|SQ$X$(^xdw m\[51@ikJHydOcf'S`x۵.Ѡ5gdLՑ:h{1a5~U#ѿl_+)W- އ 3s&8~'/0NR޲ӝ,%w[Dɭ?T>`"Kw $i$9&#F#G̲$d g e悘4‘4H 6sP.y9J\Ɉ7P\*OiU^2K;PkGuw)Jb]ͷ guMy~Yl_?Egq d–wmH_a>4骶ljk+_Ra0gR0ΨRDΣht7].W?wNVfϳEՙ"K5IAT!Nb֝?I)Imb@#3)JH3v-5$ S( T*pVTPNjV*k[`xג*cr^1|NI`Qc UV\IrE *# 0^3*I)oG)*]볰+zi?&fLM{$'mwQ9}͒3lE˿~q5/OOabJ`ZϾ Չ_nۼN&$ܰqϹY}F쌽jn//m %@=8*4D)p?D?[BptXjJ:>2Xպv;i] V*XK+D2fT݀FnqrH3A*Ƥ~S)r*aFe{d @ЉCH>[X?G\$niXmE9H{N*LIZE vtmWwal^{ pG;u}R$A5%rK0r(8#JW7g9z²Lˊ~}\ }G]dtKr%VN\vg:\`Q_EۻQuy%<mYmncJŔ,o/W WF ~p>$s"-e5<#Rvbz#<&(y\Fe,3Z`ju*MjL@Yo֜މߓi)L_SCbxhSCZm=E=dȏm6KoK'5-$oZ+G1O!We$rOʟn4 q=,у 'ېwdΘF(֜%-~yz,Z48m$J4%H/88hDg9΢҄$T+LPJ τ`Ia4ށIPYeкTE欭Oպ R=UH:mxdCigȰ*Z&F׻ \ΘW!kթHK"p9I 0 ԡW:QJ3IZP͹]6RYU]A uގ29TS'jR ݤ#&݊0lS ոZ%3ȋbk FT+oƵ絆F `zI5h~Fr=Mժh*2EsIA?Y9ՂGwH-H)Ts,$hajQ??$n>T#'6ZYzHO p +k-9E:#ւj66ˈc5jokFFGbddS%G2RK[=X_?XqaÁK?D9~hOS:ߤhl^ߖ9%+yg^z5Tj8[GQ׆@K B8e.IA^_}έ&f]7-BѼKlQ ʶ#R Z~.oqxZv7Ofus?5#kI8ZA,JMs 9nř-,_y'?_c饌ZˀgRSmCLt;'5tUAP'(k<7ω]\>O2ND~aL'`X,@"J<' ๣ fdOR V-IPB$1ZKMw M홏_oGu1GSCDW\I=@[E@͘P6֑#tM εrqD9LgԪX)%g@K&6A Pe+3 ;. .W* qAfm217J*g-ѩQ,Ɍ`1ZP'PK%'peIʄmC^":i˖γ'C^0o}_gœߚܞ^_On*߁z1,au&FZQ=z$Qu iGϴ.Tz"հo'%r-h; {% {*d]ߞx?YaM5דlyu1J:$E/V>fTmVwyS(pLl4VY+5ҽr }xs[RQ}ؽOHFat>ЖzFFj@!RR9(i,siT3ߞ)T}z<P=o-S"[iJ: Ndƹ Saw!sY c^5X޷=VӦ(f2҄siM#pC/v侜ؐm,CyP ؗ `˿9䗀^%䊞 Øݜ{i+yv3YϳgA?~].\g=FifDjo2ͬ4w) EŸɿxRxI~Akf>B}IkM{D}{n{(Bj`q*A| inY&R^T] .T•c[kˋ#Heq\{/2DoT8<ޙ_.Dϕ7+,B`źxdU;&ܤcˇkpoiH0K6tX:s%u&KhÇ0ֿ moG3 JCD{y$$_~Z=JB̉KU+EͿ}þ{Y yϊ)RT0^;|ϿZ^zp.J'rrZM9rj01SJM\B'ӓ?Eb/bbƃ!mRZw5ڬU$j@D;>c|}tQqw>,nVyED~ΎϘrPn8YaLN^hQuzS5]FnQR`k]OZZY^Lyt-&xm47ћ1սv2:ڌ!QƐrfLԝqYJ Bo,p;}U|m60"=ܽ;,{ShquA箲z,/&:vv P5u{e X35(1h\ets%FCB2+(#҇ ;ͩ/l ң oR-%NOyrMpZqNUfk[Z ^N'{:/Mݍ۫󴨑 '|[Ǔyo}Cuyf5e˔obZ{CrauL-_󵿽߳J,v:Ґ?ҩ=я*$n-y:-Q˺2jZƋnh\E[|d<[[ĨNwԱn"ͺ$n]h\E7t*jy$!.ԑ'8"O6S1Qk0ҴةkY1'w*UQkDJQXk{5Dj%r"5 Y#Y*?kR#8{#d\θ`R6\4թJ9CHS@ 4!TPõH hyfRc~J=gr%!8Z 2 )珨b%Fn <oxUI?w֊x5ˊ}?yWPW]3gܞG^/IS~z ~{0%ǫ]<_|:NtzrGUed#6WOdt2G9p`x5篖:l٫qUg`]p#65 SLo  ɲO Sv]P7ҖmrW6R*ة(kЫ𦝷t4Z48"0aͩV$^lP\.9MsiZNv>89, :p ELɷ݀qvkA}Gv:kMO n]Hg.md nCnm1﨣NE8Fᶯ n]Hg.2MqCaaWχEo|Vst6[cčQ*RpMRegƝ,%ĢllO~=\xtGoqzqbt濬bwfab[@+eӶ@^O?Y!xڢ@1R(5{8?3T4F\k<6AI3dl\:e,%mp( Q(ZADPѨ\TD*QqhTjx秢l6AVH&Q$F1ꐄHibXGGQfy"b cTFObRxlc

>s?ZE86n[2%KsI)$oag4V 6/kP% uy*ٹ8S:N@+C⒄<.I2+ b^G #%_p?M{qEC>W#S^4UΠ'| C|I :Rt䢑9BS#qyx̉pr 'U [?E1IbDB9f$%b0J%IQnjA`>έ^_AK`µ2gV\H ApBLk4N!WAEns)3dLb0f E 1@vHe12ȕYgye, s&K0HHH2b=*g[)EihvDuUT٢Z¥YUT"!P1̜.Bh wBo̙&юJdV,N3Z>F*20b:Feِ`erR8iDC8I 2IP̙`48DQG 0\w H\6* %iYshi 4ę*o:?-BMI]شB|NXtD~*7¦܏&zoLѮXiOQ㡂9;Dl%Kwc[")$m)k-l5`nU_{61H(dR9l&V, obnlb 56Gn]7v{ is!%61/&jnbgjs?IOz hBMֻt\xx-!0ی`^} 땣벗_Wx/g|1׸يonQz_Yr\pl,npb9t1򯜤 #(ݶ5^ nZc^Jע]To]_JVNxGJ<BBH~, nD;& #9 Uuuyl9WGe{RX ;R(uX6r4v=5nK=w=mz6O_\_vDu.. 516hg|mx a!6J!11e 0hО(d0:93ZC5>g_LߝTґ|:$Y.YdX!xP}+gohnvr_~ϊ1*_c|tl-wf?6P{9ӽsq78J߆MߑIe>ida@~t?ũ8KK9?_58AGw& TRjKun羘|=d3u0&H:y*e.C [ps2N|9Y\'lO,_o~<9fZB6 EñS4'٤ڝX;A|\zXGl~ =9,C1y{YM/G3]|镤:.3e+g+Q=e~a() znټTӍ'nWXC]XeC'^ot-?"|holV$362Iݴo6i:ցfX~>vk_ݺ\D[ɔ48dP Dtھv;;޴[@օ|"Xo[NOzqzu^z]o۵ȇj&E=ċw/?ijIrz ^{m 8@J亣|_tgcޙ;fc4D lQkrX'km7Zy [,qNΒw_Rx`b|Y| ;XRm|b<'ôbJtQ9gZ;2^8=mP(*6,ҕ ɥ~3zsGҗ}ߛ<:~(}i,ƛ_]W+iE;oBwYlQ<۩/Kc/[M|=El:]GYkgAtom*I*~FBlj"W4v َ*-T1cP xfVs #WZu2<t駷L=ʠKE5Ȏ/›F!P \D ftrS?mAW*.*8˧#ýً^b=Ǎ_`B)rC;-O*z[/^>o|V:)OӾhs'&ƞ*IEQBs]ߓܶWTz%hU~J%UN%+*+y-8WHlp:\p>t7 " 19F 7귴6h}j~XM;n^n괹ژ;잮|As>yW`ty X X&h/zLewpʟp3v*2ȘʂvQh1D?Xo)reTIS 7A2=nVy3Z k7z@vFvoD{u4%C_ cݎ"Ua[P1\yCXAPD-0L2j'|g0×9I g)i<cqz6V= G% ;;[ ڏp  tS%qI'"!?Yho zx&V~"Dr wbVӳk8IX9zi #Y9EX9xg5ÿ[8GH14 @5hJ,=B!`D]$Nh ƴü&V9bb!#6X(kX$A |dƗyi1Vb`q;#}".ǛG<IplfȢs70Z!2M]p(ovul Y E"C^9X=(3J80q5kw,NiM~O/z_4ztc ZPlvbӅUIZOw섀] <}sw3wUk4Q3 ![jON ̛3mmw]ʇc4J ʸ7eC}Emҗ5e1(0Ρf-Cհe䚖e̲WZe9S9̲\HQQ S)`d]bt*SZyU'l`@+fljDC2wvH{m"cbOTYs#P{V@ߏql`Ԟqdm,s9 $g :-VЇ25@BezZe1ϡ6da3{LwC\iXyԷ]Z=ei+,m(",.j ,de᳌YVpeeZm\ Ru_ץwZL;|6?&9sY'{=|?]Pz@g~ܼݧvr]v_g7߀v~ ;lrĢ]~1>:|fC_eCiF2,T!rԿ^GRjZ+Uy56| 3yʢtU'GLK >.3ޙERB+mPbnWBӽ)5.go'YHEosOM:dZA9W[)-2fOԒ _vD?1zHCco/[ɫמH{kkΆ ܄pEu.pn}(2b3_ӷ'vɽkJ<=A{0j6ڻQu0I_~ AmʕEfXY>$?Hy UK )EVPpjA [Z@\U!*$?yb,ȅ5ȭ.tUx+oTD `ErT>]_/<+E9Ix0ܠ&t+ {ND{4/ΣvUZ\="AvEX9Է] fY[loO,nY[e8Zjh@fY jcZa,̲!Sv>̲Z,ehynΝLG~s .ZO 0z^+pϭ9/G0[=cg~F%A8oaWɴY-$f6#៏o9IO\UEcBΙG%MwTż2XĪ1,;[t"kIN+ήX|X4Jב ˮl-vDNBSgxSprWGZ)}Ѣѣ58Pn-NjS b;Y=,˜jF1FMN52Ѓjp(ŗțO>yvSԹ'HL6c&3Ō 2i1ŒKpFo.RA}A͠Z#?E#{\ InYyQf?EJT!eGrd9ڴ;Uf1" IB*G"W=l7rBi3C*SB6=ɩH}c5ƈZn'7^ce=GG [ jv^cyNeBqxd1n*X3 (&v 0jEL&;nQ-: FxK78D){A]-k6beF׫sLE6"3c^E@('6`DT*AeSV˘PvQwp{75:ڶ 5exsJw[7w@>k~Q6vp%InWDmnʕnJr$ca[[JdAm׍XeNlY,AMEz~"2B |]ŋt?!VHOHka yJu?>iaɿrq%=+O֞xBrXP\F_a[jHZF}Od{>`C`X 5V9w3+/_!ϐ/}YHaA,&rv;-YIU7%!%ˣ]2q%ȀM.\fnHie/) `ꋍpK> jo^s,kT #I!/-'c+FWA!b.z+rSȡ:L1OB9 YVF%rLֆ^UT̑X;e ,sfcu֕5}lH; q=V,(?m޵-Ӯo6h,F0ww} ="?z;ޙ1ߗ͏~4o3JLa^~^<;cӪg*fSOzQ)}?v/g?ynP;,:wһ{_Vn}Mkarw{΢F]wBÄ5M #KywzÊ HN-D,J MҜ0Wؗ6[fE&c4ۧw42׸=,˜&FuNN1dCK8Vv˸L#DRyj ѣc!%/6 Dl06_=;y6NS-{O 1/\QrL=L(*-"07 lP'i9XDyԷ]iGR'oZS˄e=l̚ZvԲ^pzwO||ܼΌGǙPƘY4qC-⌺XMbf&[xk sfu@?x_]3I(Ui* [DEJC,5XEtdt}pޘ^ D=ZrԷ]4Sa_aJn!RVZZ]k?@ea#l7һCe5l^h 6ZC/uF o ch Fl3grA0NQM(VV$<<Á_&1CڠR~^Jkb-D^&#RX)ċs{mQڲ[HO)Z>k`W߀kOF;:Dx!Q{oYkfe%\%ܡ&^Ph[CˈPh[Cq?$eQÎi"9Է]pV7 jf{'Ky|Rߗŏ~?0kg;3 z()XHjk&妣,KSl4ѱL;O;`A;|=ޯ=( \HA"fTytJf'c_{\Xe+Ufg$hy ?Pvɣ=Yq^'oʱSNޗ^ċ0VPх:BXӳ Čԟ:ةz> \ u]ڨо6#/d5) kmH /GU.l!mΆm"KZ7~34>G2WUxZGL馘֞<ϕ6)`n,2h- *嬒LEVK94 L4\t.0?(33#Tx1d|,$ϩ˥C~) rP!ͫf7_C/\&9Dn{P pLïΛśb8{Kx"YuJ_̅b9 qfRQϘ  7Ew:sv75Tͺ%N v\ݨ8d> F8AP6ȉ17U1 8$5k0DՄP+Z{ލ}762/`7v-Fx$VT:bjO~ϻnCI >cZ%Հ4"#&z#y)g'ȩʤKF$2x46Fb~ȵCD f$FJ8H$3QǴp<2s2Šc Qeg2NHɅ 1ŋr:rpn' oLy\ 5NG." #/a,timP:m:rJ)S<:H4m:$x;lmb«$k0^>Jx7~^VݔPcw[ !I;Wa'2T~f Sd~ZD]$&Vc5Y0^[/JB B }mnqoDA3"X jSR `/W6@N\iW=! V%F H>/ff&@=CM@bUOJNW/~iFDݒ6ͭ$ase?1W}zېZܞ'"12 a:&YwQ 's;A'ĤqWg >nlF!3xfe~qS~Pt͢5,F]ewߐ#>zD߾~ӳHwf156ZXq2boy}ShVR^1 *үG$O}{oqH+[Y|1gE?]_-?qecs9uTWBZ8#:oߟW<8t,~ڏG@n0'$HSs$(éDp\v>jv*f\Kɬ,Zb8YS}nHZgd3ɬ3/PjERھ<]>~\c:Tq;̡H$x5jRvmپ D|rG$-!@{$'Oњu #Ikى<x.Re BΑ2Ȅ19wkPFTOj/ Jt#;w{uq*xS$ !tME?=FR">ʳOt~? *'zR ??@H {%|Z`W377_7G6b@/w 7+ْaٺD<*N{_&yѱ竻g6_G4\݅'(1<Ώ\@dfш><z1䏮oNM kѦkBյZ/b0(KÐj@!86pSg6N&8Q_o /Bfn/T%sf{#AȚt~3de#RO,s5ZGX9?^_ْ6?W=߿:lk1^^w˰P<ة?euTE~"_dJş/x+xLֶ|XOPD+Nq7,@_D4Ww\>8ʧqYBbNǠX+n ;4t?:bfgWl곈Y'-]/x=Dbǫ/Ҽ(x'&^QR2!{o݈XF)nkѺ&I泯vVΩg/}M<Ş}F5Ӧ=yr~L3Ǡm>6"ƙD)/+ޘM7ggh0^dT`ؙuEf@n~6GU\,7TGP-$>/f N)r46Z Q7=oj`ݽ/ ќ/ M&\CGLnked4L-#Io>zW4aOGCq+8_|{.(݀J3!^s_3p E M {zX;lzuZ#%Z{ tb"%jx8-=tn?kV ę N}/+I_Kh$eY/e$>W)AiE5sYJ8;"WVB{.[}JFTg)D:Aa8oimA[ 8>v^;IETZPe5q0jbAm\X@ŞW๡iAt=q!^˂A7\ #(#i yP@ )͢~J-OW<55U U~DE!u>Ϟڲ(hQa7?B𵽹4xȽoA̮]zfPނ|cy&kVpSMN*7Gf {KnQaeyّ|"F ꠬k7AhNw4nGt:O%M^hvBBqm%SnCu״ctڭ-)MPGZ/4Uu!!߸d*D,Z` W$V1R"?L,q\CzM9hao[wߙwXBq La @^`n' lPս_&&D(`= ش!g¶\1: v 3*u1pi&|p /uZ>;7Uk:G9Xi?2??.*?@}Ӂ^/ys_7蠀Py@Nnab%Lv+5{ g퀌lwQ}Qjvϱ`)U@MDi=+zuЦtz 0ɲs!ɗ<"zB1S8c3gLhB;`KEb*c۵Tv˦|¶|~'TeR {./֨Rw޽%J <ՏcS \W2]V\%;$UkwVo8MLymڇ"x}~U֜?{#3O?^.2*zx] {u_LlV|hɇ-lb`KXc ׸UN +-"62}XlpR>)ʍw}6`+Ta4_:K(1]9w#rPL+aWR,Jצ 8"Kz5Ƕ":NԺnZ_fJXM10cQf*;;sqSgԡʐA!YG3K9>}COURR m >fB}Ro=B'o .7YLp':ͬ8bKZp;: I &j'4}R>ܦq/sT,4P) ~2hˆTv7n}WnUO%_Gi Dm']%PM'#!V8n)EB']\{Bݱ*ͻaitӓ4])lK4qatBC2|Hx"щNHe\\1r4V<U13m\X.x8IC$'Dsygj[#0B;ӂuI>-qr}@= ͝F,F^Hۜ3I<@؈\Fz'<[G8;ݚ

`~X1rgm ke/4j +᭽V^Ռs T;7&}rtLF'rw mqٷ]l+g4U}Z&ǚ#$Gק>$R` Md.ExVr-^ 7,OPESEwLjڼ]kͼ qq0!uj >F,BDw:\&.c?4 qL2a %gpc!9twwJ¢TUF5D F 8T_Q-;է?^i|;=fi_a;iIXb+:IJaXb\A52!NwXt=jD5JB9OՂ?F?Ԅ`ľ: (挞B[!|%*B;Y3s'.ꗈx/hRu5aH~Kӷx5ju c/G?{Wȍ0_dVi{ 2ȗlv{ېGdՖd3j5UbUc ӔESP Z!^j'Iƣ2[1IsDka RDR%uJ?ĭJ6z̫PѡFB0"VO3>W3J@*em/(1eT%gi4F_iwH tJIƓ7B5=Klb8/ A]D8󛵜<~ ڟAs! Dct{׹`9YY‘&ϝL"Fp]F 9j`Xc4c$bK g:ȬD)L1f ATQ.m.I)C+9ϸ \( .zjs/I_f˹*uB 0ʭR1',J뢊䢚s?F1us)NelHiFw;/n#(컽ۺȬ^l((ʳv7`-91AG#%I-xƹNu"?6ݦ~\XQ7!N`a} P6ۧgْK%[>7sJ{1+?<|J*_~ʏ~9wO2a<BV ?__{۾;# t*,߼f~^}jxnw˲˫>f?|5GHQr# 0Em=~"!]IljEWb:YJž<`q@G&N$*'BlA>|mT[Ϩ|> } q/{6VLCm(~2zBm"Q{p5Wیc^מ\u_4:s NCAg  @0개9kSϒz<kĂ9D_COf,ղ\,T-T:K=Պ4&oO)w^Ɠ0OM_R)F#M~-jTW*\2xi[m mC8Ά.⽦w]Xne:_T"ޛʃysQgc<_nU{Cdlja/" ˀ^ rwN<~ezmGs(74W.:Um4}u[WĨN1X|//S>ns1#[h'"̺4wƃ|N^5@n]+}hM"3)yݓnuOyIEa栀HlPwuS,ӷ6e]aNM&_ؽ8NN- }bԵJ^scHQHn)< UY-̋.B!0c $byvJeP(}NTiur{6H9"rC0 cd*@Kq>DWPJֿwmI7o+ٞ+˦(2࿼]@ Eb?}_tG-Vi߆A[hO8zO`EjsT x,ēL3%jo̽a2SɁ12e_f%K6f;6A\* '4r/n?~zmp-|ftaյL2˸Ko}A?/~^ vyuhk?1#xطpFY-@~vF4O%E@g9H-nkχR O[rݭ :lILك2q^r[⫼5HbZh/2Rw+clV KW.Vz`$qr%[;^V. )²NqĕqRJ`X&%+fk7eB:Ui7G3;z$"(JhA`bߋCUOUynVuv')hMOqB8oޞޓƶgbtimQe-z84vCF/^R)Q3.5ދ],M1G2{1SњU@/K>S3`ÐI_MIKU$%S@;NN0ܜFkY F* INDjxe%;_CUM&OtΫTqUbԿT#xFT48Q uj)+[Ԑ&`L%C}D9 G%OHZאQ K2ILL܆(= vuc/v9JV%=p\, Gzrс1q1ջ4vZPoфy5=۫N,x۵S^J}~I"$R#()u<?ӫ<.11jUԧi0gȯMt2)f:@* LrbY#1X O^wXdIQ{1+7CTy62ejѦi%g6\24:*Lisٯ|"3X!$*F[F5a!(1^TRTim{Ay\࿡CU=7 0ס4j4yZ36Q]x$yrfbOǨ^ԩܘsӫ.}jZi_ɞQ8".Iڌ)$S~B9e{c:,v@WϞks. C2Zo)KQHئѬaϭ-&K'9& ~n סQ^}{ޟs 65$& ~>G޽/h#P}]cϼ79<@z>3030Q& wEj 9An9q2BO>FN ~1҂Z&3ЈNӖ #/Pq^;hq]Өk׳^Nܻf;? ^nZ`\6c&q^*שk$G'O@T7l^vm-d5{m&-$pU>uROFOM$_fI$^OGkֹjV܌!pfݠ2uX2'f͜+(t?wJ'hM_CTfPlװ4W.:Em;S?nq֭+bTMxW[Hև|*E8{ϭRl ֭+bTMۘQ fݺEZ>4W: ו~ 㮞!Ƽ"!\Me˄QS-P9y_2ǩ]ԩ^/0 ׏eϫڵx E".:HL'TÉDdVA'TGX'LBVLm.c!*tZm_\;RU+O}S]z=+@oP ~ѫ4oB-IX7F!}y[zO&6eN2'%>_ޮ;r6Bj/p3^W/B_<Ք9Hy"Hb9% @EjyG/q|3ަ:6AICub;˨8#.O܍,V!$Y@7#Z' mܷ,U^fiWx+j /O~^ y󲪟7m"O`&sf-fXB' #R[ ÖRPܢӟ{P&uٶH0Q+H ݶ4c)΃VhG@H"lVLH+wC ?<|T`EqcXBIW VLsm̤JG90la5e| әu=H&#/FLxZBͪOEh2;*[V\ٵ?67?Bq9{7yor8=hL< <QnHD ;䠋 䤟d# Y)U 7)|z1E= C_kI$XyI\.Z&’Ndd.a8m>^.:pxLDqus+y.IJ44 yfTKT眨5Gl^ j.&Y\iVI5?^JssRaKDq &Q1)Y&KSo rʙ,2B%cp)LF5EX,#XE$v|z= U+[W,quY 4 ʦ$4 dfTs-f6ol'g{D͊öR7Cq֬5K8bS&YǨ^ԩRY׬8h鏲ջ/Wm~ TNɽ_]_(͆I&ɐ&1TY#_wbNL8V N2{J_{I {z_VR."'ccOM?RnjrJSJX@@Rym >Ee@5ZX⛫gArA2lmNQ|T+-MDOQ7Z|ȕ$liJPzHJEzi% $dBBJi3՘"TiC7;zL3I0g)s)3,M8v <=Brfǣ7&ӛ&<Ӹi"z8dMM$$ &1xɨCI7hY.{Y_r"8mk=&ٱI7&F ttzƋ[`XI:T).k+T/i۰v4yRw G_.сnA7û&2@AUJn~jHRN=NgㇷfxX|t?s9ԬcX-\kHco굿עU@) XYa"VQDKQB~*8<İ֚`QE*NIU^ぎ6 RF8DUt  E5e&sS!>hEZ}Ae鹺:z6VϚ5fIYyM8wLyjӬ5k,TtSĺ"uemL̘ĆE<IRylJG24 k\pҨ4 7s g|͢L'ݍ[I)$MIZ:$e`SJ'jscR,M#63ڰKy3u4u9Սq xSPM[,@BcI "XFZWqk k݈Fx_X#wu"ZA 0>1K2YrKxiв|c=xfߜpjP k2#7Ͳ9H615u q\\ EZ=d>V*H|N4IE2JTFIU8QAZO9UZ˾1[+=o 1LZ73ecLB؅mI$*.=@ x(U:@QUnζa&qc7 =0)VE"GkX;],OfhÙݰ~zӰg7 kY6aDYih䷇jXutiX2:ۑ @)G*f"Ӳ,?gaegD =1Jؗ'l)Vv `G>e 0CF O[*%vP7&6{u}g>wX jx-@2f^~HEQ_ ?__,KeX"bɎKWN8G(6<;Sb4Q8Jп1KD{T8޷-D#]$_*(u=6׹o-|5sUl9fa,.Іǹzq4*a{Y%c<|]~{d +hM֙땔yE,Ŀ!ߕR=UG6m)lw|;F tu ~|j_^?Xƿ>/ҞH{7/pwѴ=嫨0JP+53TVU ( 0T]H⻴ok[{Mu}wo붝 .ukH}ѺUYtjk MXقlE)-!*mp,D)  h;]: XҨ3@J)nNjg1u$ឬՐ7c1YT]&fH![S=#PG_*]4C`!CeJUV3 >< z*:KIòX.( '_AUyNTE@]@U1ܒ?~L͂U<4S CM,ji)m+(/v8UB,hHm6XCDvVֆEjкq\n]wejӋa_:Wj|?_?\Y#,q2??>ă -_GN3B-?g~ekڰh]1g psߕ}QI}VZrl}:R"22342Đ "[$a,}\6ga('WBf|f <\I2R:QG:]+OwMuYKϘ?q~m`3>ٻhxIC˛3,dv !tz}ɄdV3G* PY׫'i$n=$,۸OuZ6cȒf) P 3ʣ$x$iBj'9fQg-ȚAU}J+>y4}cu4m&mZ?S]X`[ Q˼:2;Z aԏݑ/X0iְ=X+':"kH‹h,׍#jB}yb#m`IJ#/A=un2?t(DX'&"}~ztUM@x$2|^*38ߨnh6F$M+4ڻ_ύSz S̲%W(Py{~7Rp#&xu * ZƯKPJ׵#>S^*Ys0v}ӗi4I$mT3qiif4R6(/Q}ަZKKSfN1R1R?2M`@ ^5V*ˡ7RJ ]Um3%(PS,ͅ^ojSd\-/S}j<8VZi\kL!j}6dmד- I?3[9j=,Y$KT&joU1$L-sKVlF*STC=OejVM5 ۶34;2N)D55o,B!4I?Ԭjih[|//O)$&aHij+y#?NdiVznR2$R/?aaejϟQr}ܿ6xEK&CDXjaxAeЗUwlؔ7 99&HiciB,dFбEβ9H6!lX5z 'lkl]Gq?w'i~^$Hy󢩟)JĊ5Be# c*6epdO9^S~pf@4Mϗp磮&VHc?7O[1ƩUTW+: f2L2 RO!rLkޤ~FYL>&t_up} (/S}ަ[|2T%-fd(Fu,K7KQ+U)ޕ1YR `PE%Uel*ve7-$lhd"a @[ウbXU_TVYrEљJm,U\gz#9r_)e5:A@? 2ٵ1/3ƺ*y=YVV֡;Kbw+ $A~T,!{<@ђ-zfu v)ٙ(|)Ȍ3׃ap;S?7g90-pU͙ Pyz {7ATmDA:Gvt©H_-&j{m6ܩ \\ rzYwR5R0SKp4Q)?Edk{Z/,]|wPRlFO/Z2-+BFZ@ u c-A| B= 7mzVUBND3,Te)Ȼ)ETAqt K~ H⏇x|M Jѻ@V FT3K LE]J`(!(lv$ÙF t@ϥE#C4M}@ЍnnK}oMzWۻIk16J:T+VfvT6t :\vlIM!a~R쉄88aH[+"N4Q*LvzDΞxtocgt5te%Sh>4(-橕" LRt 鬤\9ىZ, `}: ֽ[3_mXxl|YW◙\r!DI-ӹ3`Ȝsfm:2+k(I43^rLiRwsX3M_- 4 i_J aa ,4H a,dS,Z;Ȟ6|uTǿ|D׸ן_?@OV+>>ŏ~Å;ʈ?vjO77n;"+zo/¹NjMZuOi ƫ#2G8sOww]~agvORSڏyشg|vV9`N!Ir tXK&pnm=Vri$+b׈| 1e=%Njjm .^eБm/*mzjj3Y2IX=>ݭ{t]N />'_\7Sa4{߿{:tܬb1ӥ4n&%zk)~9F6MQ#M冘껹m~yb樶MAk~+~q g[|p+xZkHԅ=_eX%`oJaPTJ&s$fg Q2HX dJ;'cQsO& 걽BDYߊj]ט`!Ɇ[ 9S-V "ZE6Zq:j0C\"'o.TK,ӺAHE.lSĖN!oxEMsU?d+5vz .^/\˳|UVɸ%Nc ~ppA|]# 7WV:u%2uk˃i:mT[j@օh+j:zx~Ϻ)^ѺAF6t@־G/кu!*ZѩQ>BMaYS!o8 E}ID )Tl`xIBfd.Z* yk=hJa` 3jMըgwUjzx^yi)YQD% q炜57rLXBX*ՁS#(4v- 4)pj8}M=RljBOw&$;Mejɔ@ kUzȎ_=M\C˰UP>7q[f7 ˭MZ+=H*,ڰlI͒xiRNQ}Gsn}x[WܺRӈr[Q9~EBK4N)RlaEXȨMǚUj&aOQ}G5ќ~ؗ3h9{V9ڲo}>}#Gg#0S~]}x*BӇv/s)^ ]>ޫvV^g?c^ī&f%$%0VB)X(68 C3dPR.utS눿R <  qm"4@OEHGrd=PQw{@2}l?kфmW(*{ZmEr_=g :g}lT!Xpkyoxmʟ :o&D-%Fcg\Qbrml|5n=C>wppp&E52j&#"۳lɆD_ /RLȹ4_|5>%ϲE}-W7~kun3?ʋʋʋ,(B%iL1:*cLIf6طS%ylŐJ' [cA߳Lb͵m?~mTG1{Ӷwk^m½ZkIJ+@FA4 $x.fj+]:GT / zg7v S5VB&{0眃;6pBeY:Äb kt>YY.akvVK%xOf2|)m)%Xk*,@Mf:]?7Iy8׍4˝}]ga[S;(^G%Ҭ=g.yzGAjƋC4c' ˇ̲KDĻ?`~dXX J7Ą2cZL2:f۹Z6 ߭u3 c،55 ##X5672aۡ|YTE2kPEYTO#1v*Պf Y3ܒQBěשRKTl۠ܧmN;F:#P0Dg(V0_XfP4 .qhMA)O@vv`\Ȟ6qpcκ_+_~݃..6P$Xo,O׷'0Q?}Xz+nn>H+2/ I ː84īw۫sQ "F{# .[ν/'+)'Ɣ{_q7}n{Gq^x Ǜdc D$Rs($JF_Gqx77xUUAڍZ۞WUXP8+kXvne !0S}YhWDj%M`rK1p#zUo/lr?{ixs&,&QÜ6p΂r5$6#P#;,5g1g| L[(5Izum%svEͧ>}j,ݑU6&bD!H8ʼn(ՔFBH3(CEXǧ!o8r4|]HUؑZ21?҈WcjL,dfx5 3'hA= B ,L+)*͚5fQ"dRf3סװ/RإPj8~)-[a uW!wGYpofz 3@tعGۼP)-߹@Vغ߳9+oֿY [cg?ӵ諒 5o}a{<TO#P5ռAqAaAL{}u=Hcz"Dy4g;C//0պPZ "Zt*$(6KY̬)@LR \*Y/hfT'S#J 6LMjgY{9+ T û9ξ[p>0zzwp%6zH×3=xWTWUat$a+M a`tZ *r#l ol QOHRK;K#ZMB (pR@Ê kkh½pxnf* JUHn _JERg9%g=)3 e:PNA+O-PkdYhyј(UN%2 @8ZAȂPJ^Dx)N ONnCfAn:,˖X,a$h'(@9S%j GQ" h)ŋ5E4ͭ'))~ x;+3O!ڟ?tsϕB۶?mA r7fԖc!kԘ%E*_BcͿHq{u $s_6Yf/ ÑDÑL"$hO8 ڨ30 )ъx:U?4 JKaPEА65@*8d7,F&$;RtՍhJ%1֑6bBu+P{NL5#pRšB8J,P<jZ/UIIQeYk=/6A :HM/( wb,̓<-wZL;$X8(@Liy"A_ E0!$uiEIB}vߧ9IFN޽/YVTDz^O:e_* uV /G ̿ݭx(:%h"#3ao:gJ؜)1I˖*9I\+9˄!1EW >$CY-, '=V[ Lqg$̾|>;h;[JTOpus~?MO$7^8=b0]ը۫8y[U8?+oo 'oO_>@IZ~TaTu\+S ]8RGu^&tAQ~=ЧEм 5.QKrV*o(H^GSq+!J?gU?E#2J-ҳ|N3@Sa QEӳzb1t &$SSҳKJa|}t$m}b ^jX66Dq6hjO:݈`nJ$0 굒LeLjn'+N"OjbE$Pƛ#aצD:w YEXAx(DW@ QP<5 . Œ8v"#TRr, kFdbὲ$x""d# ٤*3Rf}/S'2O`p`gTI6sŠ$Bg('R>Ix*[mjQQrO lps/Q'L>ԋ޿?RP" 96=cpqǴXJK:! #nIb p{(Wc>[f^so7[ןYNV>ϺX/}lu3jiJo)y{q|tDF'KQr0۵trϣr =_J%HdI=C3m[չ芮l/Ks6nG~ǯ~r?k&Nj'}'B-< Z'Ea>(b _u|i:7“z>q^nn/ajη..pԈs pXA-$pC;TDj&$) )I(&l+)i\#mm(>:˔(7t(q6.̩Smw0Ǿ~baOT {Ç2O?_:{jJ4IRZ>;QR*afa:@lz_nnCn^ݙm7g>jIk_D).zw @ƉRxkYS#$};'E9+2=i5Q閊2{X{Ds6Ԋa X*a3 ϣUxcȣ*0wwG ?U `4涖0@>_l)EQ 7坟W?B/e8 iJ9^Y׌t\rLh\WAv4+<-NV})Ž8g,5ɮt7J\N7Hޤ[D|,S-]fj [.RX'Mې^&zn[OD2%zBc^=h?_1Υ,xd==n},pverY)P~m1ֿzp ۫/vv=)W^}R .I` <&k wX)a;b ՜B#NMå&)ż0x.b`b2ev#9P_(LLf# fBy06**m &C9$HdHA50ӶSt ]I3P(bF/  +HPst)2puoU-Wmfmto{z9 q(GmyiR6Ωd<9QKYduȔt&{%2(%=uKIg)8g⵺b~"Ժφfj!!&-x9oF?f/?#dza?is}MQUq=o2z98s;-h*sгF-xe\ܕ_ j$#W?< Qߩ܋Hf( xDžhkIi&H.ҫN7EE\A:-d&T. [m߽{wP 3{mwc7w1<﷡6C?ڏK< tf' od`^ތ^_e}ᗟ}|~sc{HRkj$F+3O,Uۑ#!ImvPCBCI.L" (m,R .:'(:'NIbx<+BkK;z=l{7׃95B#wup{(&Lu3[֙}o<Ԛ2}ldM_ۘqY}IB,R:>>v |_?e]mЯk?YUYkF{Ln|c{2?WAs;.̅2,nC9Zu;Z>0LG[5gwSzEN"H9%!c2N( /7ڼ`ūoWcǦ@e*.j4A~Uم6(ԭih0lRM7c:ɇGX5ތDN~>L*T>4LUhTDn$I:)KCW+5Ӹ܃:*X+r4UX$.nN%iG%8?P߁ZK)Piu=K,6AbSIl7-T+R0&*ATYc%xHʨNx JJwhTch2ܤltnK(D>AgeQ;c (ulJZ;ӌg{SN_3P{$&7-^UIé@K0:iBJDI"tm^&m5zeP_ܼoD7k5wQ3G70_[W-K+t %Jv#h\^hHae6|QL*%T2}iRY ,Kq4!rYO `h }xlL9-3̡el9-"۹LB)^^P ^VD}4O^?C1AbV*5 Rȳ%l|\*= f*Ǿ|4:e!s6j^!-faYȜ +F0ͲVPwr4*XNU\ )"H**-Du29u9xӨMԠ2>;I<˒B*,I<˪Pe߲xbnͅI VȦF K .UJi-,\fNjvqR^}PYXk\,k|Bnӈ$~n'dWY&'iL]OJO^7QPK]a&XjbFfA#Z̐-&_k)pxSA 'ZN[Hhad`ƶ tu-*A8v-aƤ+oR;+h_Ɍ1B*LpX%R@&$$Q"9W Ҭerץ iHH(,jD")h\tM7 8T˜',W€#3Z#ÜoM F`H<ͼDPeƘ NsRYu2-W9`5xJ)XcIp'LtNI֧a`ss[VO]d)pB:TG+Xk+2M|8vL! jgab-V6˛KOEYu"`mw`kk;ҨP`mԗMaVVR(0 b\q58f<\` DB[7ϖ-OmCkG[&`҄ʁ ּ˲]`:MH<1,3Z.Xł5N2W*8ӨM`aII U1v9R/sR \9)uh[ 15Σ`nHO 6SOb}Q3L!G8pҌd| }+滧V}܉;h'cW׉q{m%d%Z=PkW|g8}>RpC Z$1=T_EKcs-. WI m!B8:cRs"ym+(*OY> >C$_OAA^oUS\J+auLTj+^>4j\6!Q?PpP@fhZ31$/Zj.Fb(pA`9n,'a386.Ba6Eꜷ%eisG6' iH<}Syֳ֓Ê ^-!(D>+*giJ% \iT .,MuFKtp,< M>R4JҠr?StA@q:w_>*H:yռ(7*>_5`"flev8pߣ&Kg0-˄A'`|DB{rZQI&JHf@g*עyhE,KOLFD.2-̦eG$|ϲPexT(r # IGISTšL&!E)íWмRn﷛4N+)?d̞PpUy(KX~Z:IO߯`8ԎMQ}O~$B'6v~^8kKf,էobdt|U\ݤ^ 7gWNQDxsu>ކ{Łt: )A?Y ~I/M֚~eU$BfI2~wWf)m$͌u[atEԽׇClS|u# !̃ÅHW79X)A zx]p]43f%- d5JePuzuzP&Z2-haYZy`^3#ڟBm-ډ(fRjnIO1vt%}W$^ĺXW/> iįuBj- mٽvWŗa ۪FI9<޿ƌqTFbo$#z&%jӃS]0ztVH3xǺQa iD#uǎg[}RH6QZoA-~  J A]MEtG.ީ=P[aG -ZmjH;uֵ V1xu޺*7w6Ǜy70 W6m+T6'27^6յ cxwz2NgnӭѢNͻscz!,nm u0w;$,ۛ"nsw7_xZݧۏ߼yKy{s?m\v<0*"G#E)&j0 GcZ (֦Ǵ҈ڀZ<ىSXoA,2s+Fi N:{b8Vgښ#ܤoۻ?lM0M$PVi&ڃ|Xr;/ŪSOUy/lT%?RPL.xmNfO/2GTʐC4LSFcm<,W :cySF6շ1/! vrEv44v4y[dt7&0rVUZn;X쏅߄f&yEM۲;s+9 QC0+O- SKiڧ L?VϦ@#xgyMZ^ < ʳq.:DG VxfE7tY!9 '>`r@/& hMrK [碔ye%lK\x# 0 :G3X4Qv.#o?, kk:{ְMT_f cP2OO1ڃ*/tؔ*¢Rcu.E`e:a_q6=0+2ў,s@V?A@Nju #S+.Dv}15YsKw0)oQ `Q_6ťB+g𘸕m#L<|/)\WY$z SP:%pUmUM٢`CGW]:LU: t)$Qv/)$BJ \1,nDX'2"u$+, Y;J$sM!,Wrv8xhH/}$Q-VUчY8%yUFKe;EL+,)&jZ,otc{Y?uX].Y)o:KN^7Q2PD$gwCMR.2FOtdqKdERk4vdm3 RoЅۃJA.ܪ7Q3OQWtڬ62.v ks9^+ i&jzڗs赃9q̃=j@7L˒e)kf$Ӳ*!֦]ĴRVZr9i,׿(²YV:<߲$Ӳ5a,fV^;G@ eϬ\n,0׿j)0-:Y3RaZNiɇl̕RCʢB N#?)xu+ Z E[ ugh]m:P*͝%pW{)QMdm7kI 6brb$WO[$%Xs@&rf.X$#@Bo,Q! K!ʲ H'9׈QY<1JOb 1Ͳ^RK+%RDxN:IL覛nq5ˌ 2'$\YslCe5NsBLKY>}Y(Vo%Zee˅ʽYx[RxAg{3B[᪥`Ʊj\1#2='1IE-x- gQ}|_ !ԑJ^-YZNTyvauđtsW}jǞT2&@g=(VC8%uNJ]&]'A/ٳR[ttyI:}ɐjML9V)*&Y YVSF,8bْ漭+ZT[]secVыp^A4鲐yI # _01/">s9IA>ͅ0DM!QN=صO2#  IE^V c [ ,W@{0DP rj:A }Nmt3J nI(1UalNP"BBC9KH9/禐w4k*H!r~^Ȭ3,rDn .O_w/*\~_¢c }%\~{7_kY7?}yE= /)V!?]__}xETKh(VYͫRn׹%2>-8Do.nPg~W!I=>'~s_? B<]Ӄļ1ؐ3HM#Uwn|=Wsu3_V7Y,M$6`o5oFvFA[_}1hpXK2#(ZֻPE1m({j`{uy #ʆO%Zk; UKuܫljVz{Eʽ櫫ͻoP=4POր]Ja襗$.(gEj'9ۉ.2ĻY8NpO1Gϲ'!`kw"G+ \+E[Z´uKKp-㧟"|sTe\#G'M%7OObB.IJYU̟k,ƭl>7L?##Skҗt !hG@x GgK4 :*8 gh h%; r `'@kHk#l|G_5C6iڒiŞ`lφKr<clSN(3/5[ˆ( DNcui긽GضޭnouW_nV1hܬڋpbC~v%!P ͻ~{p \`7)ŗ7DxըSmӯMI#*VNR}UW/UK1k\d!_)_Uл t~w;]4ïf!#y9m MtM}Ĺ 9xotn'EN&nrhæMҎj*M: miJsĉRU7X/^/XXQ;QR_ a 4Tb@W\(Rafۀl,K#k'CTԷZ 㕊R/'cy'4Y$9vfO`K}ٔZ;Șm!Ѳ4ЫN.1 q.TRk{f=\D˪z)3Rv -3=z!6-!>^)SU *'2gApnWs0JW }*Ma/LHAbᖸjXlr^1>y#5^Q%M,<,K%VRD]Ζ#J<3B3,JtQI-eMlY؝'>F#x;ښ'Hʳ1 Eғui=z|j1zSY=D)S(UbkN=S p*wy2Yx8N=G3qA?FPZMy 2j7r6u6\~S٤dhOi/zBOg$A/E\<:]lnnQv$G;~l O$篟~K% *l B_/Zq_-JM> ALଣ~:Rj@C8~M/KTCd R p'_l I8ՂX)⻇3r4 BߗxN\T^e>a,o9Fz" PYC>?DP͢P*5 tT^0.WAz)0Efq>.%y/cwUX1n+l]lSTPhڦb KAdd؉ݸ]KY)kCQܽ]c|T"j)G* 1O!s]Z 2CŠe%(9$/%-uNc"{Xr~ϰ"8!թL"ڑx-ƍj!F 9KYH2Ҁ{ZU$iT[U)ȯT}T|)S8yn(Iqhk qezݜݘNQ/cO519-C!$ܾwDҶLwX!W*dv[^4׸Kz8n#(p4k="ӚttADpRr,kXqF߀tј+UZWU?>$dYDQC85S*Yw.0D2]`ZwX%A (DʪEfr%u(U)o8X%db hOK}ٔWGyN@<Mbe1Cü5-g_)MޘY5$|ܦSHkaܙ$!NA[֗pp1fkS^XBYɿz mjc&'{(V֕C7sU@da34hyt%FjjS{ AQ6&~ cO (trQ{hQh2 nH,X9u )sNDw3Yջ}V{=>J[^IUx=~?_E[N;:ɵxXhnzZu|nr-6~FJ 8~U:ˎ(;I l L D-Н l 7q$2ƾꆡsñ:ahacKjȝ|[jj]e۰O o>DKm8:/ |()m~K'yUe@0 |x | MtMgRy73N݆bb:CŻ0b6ӻ OrLncXWnCl >n P RLg7xS"dmx]JwBrmTb7M- =?ϖ~6LD4fdtZD}$cXX(ZY7 RfYvg˚ʲ(0`3*1B_׎j?nye)]˦ky#r;yrN<})OvS_L!8TQឭR+?Rd0u]ǣhYLzCk(?),j@4J%م,"3sNy{aj)ftKb8<IoS$|+J=KxZ$((y Q_â&ҵb杜2r._WrƧ|2^̯AGf< +~mM\+1tdN&0:wE0A2+(ʖ2( ԉyKVZ{ta5Y#q9P,BȐ  (Q(D9hD|YPYgIlGhknIGpϕt !'^{wӥ-,")1XoPf"C< 3iMhwW=tPOS.3Pȕ~,^{ixkxLXԯ@AL(Iރ'SbYZQUEF D3Q(U{`,rK%SrOI}ٔZ:P-ˠFNRy'T-A.z'|'D 񹥉DxxUx-,sW1܂s猳Ӆk#Key!#նzaZ[ (tjUhQ4 vmhV%)IJݨQ[`)$9Ӂ6*xԂc)M =rBBф=licl_իT,J?{Wm/*I |BlJZ993f$vgRowX$zbػ6()Gz \5w԰ة=6q<<ܿ|3'ndP-[ngR#z=x'Ϣr2B E*Mxh5]"' wUfZ1at<,"I7I!j-X\eTEb*><7di:۾uμY:[jșwPCkǻ`.M죚O -$3-SKҬ'bh-+$wat4k6]y(w(qnwet4w'$@gcz:c%aw }7,9Z ɮ*@o 9t1Qu -@žxv/|p%iBY),Ha ^ńz&̔Zίbz?B,ͥf?H5 uXXN'+Ш\R~,{]c;n:qotntV3͌nCxE|Jq>nh0 uJ%`#.3 !sS`K10ep^%7l +a~$i/P/I}.5T(VPy*g([[gMY27JYó^l]j.q[Q&<ܥ7TDmJ&Gl(mTKPN וuՕ $+ijߨ374iڶ6܌K@W>+ww[H|[ #%Gjei֤%0(w~Hi?K+=6D1fFQfF3fAwW Lp}ף*xC  oΆ˸.2@q]VCb]X,\HokPպ',ʕgh aTU IW⮮ʅ' W@p%.> |L(Mし֙|G}2ʊ\Tx<ׁ|=5mZr>\s 6g#!#Sl6h C Iz O2^d˗Dn r= f4{[9*ajWqAra[ 5P\x,፴Z9&0{M[' "^5tTBMWZ(Ua1UDP/K}.5 ҇N(Ӱ[ebALF˱ ].@F 1]}%f7/^v&-䬇sb |-%3-H5e7O:Y*wRN&2;)MLYxtqwk[ N=^N󇢁R_uĕvh"p\pM=Vh ϻR MNu" .]ܛF\3Ԝ7l)7y&iV{q&H3;S?҈H  @^y y؋P ١ဗ>b"&P$Ez.Y{\? a҇E"I\7qX3n_>Y#ڕ*dk:>ӷ`9p" h8fu"1guֽnqpc&. TH&Y4WAIߵYr)\ @ה˔!Z.en(9 Ix@s3@A%Zim_mn*dC27pz&_i8i/GMKm9zFBpR>E a  r٤ksqzDRWscQY&jGc 8*k@2h/Q5BD11 5)_]s &;x , "T9p&`H,]@-kx'}/W>_O$ٺFp:ߵ;.ж{(5Bje`^ffdXk@r6gr޲IpZ ߲IꭶOK+zfwSHm\}5i)nڲ/Lq-5UcXm]ej/*)-4 զ<˞)%yQcP4eOK>)B[vү2{w@&LW['F;̎,ha`޳ìd[Hq!R w,bH4n/H㚴d76?T\T],+Nܡ@4e<-<}~t-ϓ[rw{q]11DWkNG R",4=6Faۣx~vz$պGNT)Ҭ&7Ò5Zur?Z)g3&X)ÛW9q*'YՄsc>cN]6r֞tg8UCwHpY}7a-LtVv~\0``I&ZdK In_ C d0)tىeuhv?7'!F<צH'+}#SO /+.B]^/?+dYp7}KVetKͣi\LRƥnqǝ< A+zh#6!ͤDńÖh8nݚ]~?ƿ3EČ -x8qhSͻcU 6O,8F./?6M'Fko}|'JGŅ貾rU[3LiJz=\]wmz_#GF`Z85_T聯Qch|}_MB8#x+"iwlEޡDZdL;8_Kf|=TV&.QVէyG%9tȡW?)_#f9Hlޑ|Vll'Y=:5{U!}f, ,Jq٧]ua﬉H'*WΤ;Le3fd߯).;v^+֎qIAz.qٿ8a\3cp2srnqC эUe&ȺJTAh 9ox@efB6\vh(FdgJMyXM[ aS()$}(v>D>y$-?(a NhURLI\˵ UT1Y'X%DOUʹ[3^:?]ǒ{ ro@h"[}Ḛ8u]4{%Cf9D׬`p~\$)i fidddpRg}:5̳๗d4h$qk=lcT6CEpҏo5q/ĘuW^e,PZӦ<3ejfkvgϤn!jlλO3=7)# LN,rR;jy٠'oX0uѕ}hXczrI kM*T\J'A>S&( #u*қs\F[6H TϰzƘЂ\(3JKA3ff&32j}،D``tsnz)eVd.Bl!w.$ٺ(Zuau2D" YͷuR`ҸZz]R+L.Dzw8v5.-tEqq{LY 3rzL_^&{DBdQ: {)XnK(l۠󯾄|}nѨ!#nNC7 a:w ^<=nc w*i(agG!hWv! .^a S_xP"Qhf3ߋ/óvf.> Yg0bc#mNJ.w+? ĝ%~x0DK[iDY! [B!kf[(sʹC9/}(W?7zw/Adc7ަLZq/ od&v y]XM=D׬Xߩ2@3.ط=\(vЅ`Xz1@:Ppڢ|o⎩V[#o~F@P3LT^zro6VfހCsNx#XMk^kk+MQM/xKT`ro& \Hs"P9AV#y޳Q eՙꙌ?Nm ty+{@qD:Pitle{e,Iʽ]N~ݱa t2ܻxzr,P|zmQt]{03Њ \ʽ9oV<* SxP${Lrp .[Nۿ$ٺ;'Lϒbf ,ire5GCW.55{@t̢z.CV~ݺ]ˡ f2gG mHף}.wQ083}7'|k}sn?]t,E?>Ysn+W}FORE"-ZKG˯IwuS7 +BRQ̥%8$ŃPgxeQMYY+󹍵g41)`;_Pg#2#cAe$>@yi Hx9hl@ވײQ9BkԒm @HR+Hh@Wfբm,dic5L3'YBwemJyp"+@r1 YV$9ۯȇrA#9&pFB^ JۯW }Khq}m;f\K><$$HQhWQؠ yV%hrǘ7<[@PaО6˳鬙{ͷY5X[AwDgzi<5GvHd;bўkk7x5=w&r}${B1L^kz%VIU6fiuQ餕:}ԃҾ^mFlCW ڕWΚm*^YEe?U|}8$lu9CDeŅ2<| K޿":);i}DNca\/כ`3EQTYgZUb{RBZP _8r3PaEcpbFKrՉl6*XKHTh̙"'o X_򹇨Xv ֎LI;ެ7e9j3\+)_}҅꧊T grJvȋRFyzxC{V?mimE ŏ1N'u3+S__DYg_Aܮn+h[# F$*E8Sa=沲6D}'zj,fZ+j@sWe*FSoܩ o'p=PE l 8)g),HRRp bw 3<-)H־=fdҪ3%yեbT809OPxcN?xw@r4k08h֓=~<ΠyFjwDc@BA#R_x/y+ukpbt;+p h)uE+$ e{ ^mc޺NLD=8cets{h/ڢ--$*dabnvٜ {g͊qM֣2onrVgסeނEVEb$ë`*q<039e{>6򟟾,[TkD30T#ʚp˧RHKPH|QCĶ'9) ?abL{0橴aHyj_ǃC o<0k; Z- 1HJV^: ߻ y$(1 +fVTlKPʌã7LHN;=03t] %t&`_9s5)ۥl7 mg,qU61(ot;bjz ï##m :0rzu35IVNDPP,Ju}*#rX{#^MiL[Ѣb}>KYZ[AĢzil:9zt8x^mq҅6ahWi.5y ^t"E<[Urseۜm 3[1 &ijPR2*¤ʰ.KMIVW#uZ8ƃ簩'sL402+}(wVͲ% ]׿Ȭ"BFɾc7jY^f֐:Ts|Ӻx,腖9YPA/v)Nֆ0<;Tw$ZM\"(Y,<?<;Z埊uq$<`q?.LJͪ!Z֎:rJ%~CZarQ뤵a;ң!玡w?Ku6[ "ijE&* ! pPxdN{Ύ0BP;E]Tuأ+rx2Xr*B2^vJxxf;r[TR1{YdU~I`ݙҤz>)ֲ{R==Wkp!2uh A (N䕔!Uڹ0O78m?y4CÑ%Y,%SQDtZrՐ3xg'5g-_ͮk]2GD)*,})fr:fWsHT Ϡ5 kq,Paa5ȩ,/Z`>Oi[[Cͱ m׼zJTM6ۼTpzo8F:1t&] 4pLHs3Chƶ~њ@VW>OP,IeE~ O G` Efl)KP*} \r #Rޝ ^@?K⿈Jj(7=q eJhT\ip<\KVp!x-QSbXy$0>5w{NyHekЦXmEZ֎ 9*V[PHsd4F~Iܛ4+;-WEYb^`y@m78vbjjvrbU=*#[@kcI?IJY]A;ho#o}b_0EX\2&sд;Gw:]Nr"HM$QGl06|w3=)|@Ŀkz*}$`wxT R5դ+R7< OI%^b؎bof 4ЧhgAiLP$Ip,P4?O= @˿)/¸Mi3K6'#XϤs,N G 4Vey-v:I^~JXIds9jj#V{T J%{f1V $!YIby1AX+#wVoN6sB|^\[3),G%*0E`1Rw酢8Nʔb>!fUX<XDB:d/.#5@k&j'?'0Asp|m$`LfkQҞ;CI{\m{hbӮJδl9Pp7=q,($Iwzx:k򷞼葒۽}0lklphU=uLX1XM|x?nzmaEsf߅fC} s!Av.ؐvQnC5% +P` h2-6ۄ>^z P]Nkm,Pl^dbɧ;=qAuASaO)6Ű8Q™ ކD" cxΎv\O"iէV(G85Y m.[S)S36h+{\yRHP$))0e#Xut-6mY|%i|^6-E k $ݡ‹Ǟ6 seJQۣ~)l~С '~b8gcwO*H,ff_ #y{Hjq7SI&NX]45(DgPu0E,:pľ ښA>Nhip X!yxiY<3<n_JŞkjԏ\(4"tN^X;JBΚD¾#꽑/ 5c˨^¹W):M>\X)jlil!-ґdXdA7nnj-,1=z+z;~YF9A\>O*L>t$`}q3bgu'-62A#t %OJ ȕ( H>ZK}(Zo`D q vb[QY miwoEܨ૨TAOĜW1űA&xGlf/95վ }Tơ틎\(/zH84J8k@N~h+(5 7} DY'@S\pItlTƂէ:{oe>ftӪOWW-Bag=R @UKKw֓YFrW~m,=gn! k \dk1/]\%ULi9b%3.Lt/ͪMgHZcx{} `-ZȌ=afpـ?H*GRd{^ɏAFnGfMG=j m ^kUL |}2''GqKtSpgo<–s_k{G՞֩Z?x1-A%N҈1qH8 4i -80ԠeXd '(u^`FwV-Q(w4\Pr |LU;h̘;Qg).+L k쀳A< Sc t|g?uF|67;Y#8A9%+,<4 ʚ!X/6f׊l,ekx{ .eQJZ ب׋d_Y6mFijCXC"X3՚=b;cQtt~ea-#%xJ~p)?LO*47?{2q QD___䏰}0}wr9T웹Bg^zLyG+g 38n{>|]keTjs&Mk.^]4ͨ/`9Hi,"dyhA9&$U{ǥ &Df(I@Q/N,PJ'>ڙH4dJDv YV);%q2Z3n`B/~{[s)y<*[K\^E/ĈM,R$bO" MJopؙ7yccU}N`#LuVIu8,A .EԘ;醤c3aSXx7rx7mMQ&(}/w.NH$Yɢ=jW4VRoJQH5bR3xKX:w% mfjNֱقW\]f?U"/*B,U۝4W-oI/yW asV\\HTgy +'_kO&/oIί8O y pD%P4[v?Ѻ=-Uu~2:sOXn.lhg!< [<7u_?/^Ё px Û+-~% U5z Ƅ Gjr/rQ`텄^h'aK1iTa(xvC=YC;c V6x6\FQh _R5W /[?\i-}yyJ_(dxg<2uEK`c,ΞDfaWaCYjA]F۶nNbxfGg*rvߕ,'f>`d-vm; 硌3ftF*ݦɗJYzS5l%-?`ԃv㤩.S6˧vV,Ý\ßHLQ|IoW%mɜ{+BgaP {H<`ْ t2D_иw;8fK}VmBž|;5lK,VmO6^!f]ê-cꑃ͗Ք/rjC5wM}ZYCϣAݖZѠ-Z+y\KIځ$) _[R(ZBQu=aFe~3L{SĐ)1j+a^i.T1K  k [s񯗻eAz5p77n x[mJB%k{v7{4eqzSjS^Z^fFAܽ{7WMJÚ6ߊN]YoG+^bZ! fy *ID*dբKS$plξTթv5>z.XJ0@wv4eqMN攥‡ U^pĴKWXrWb h"/:"1慨[3;"jQ{vf+4HNvc\(F ܍U IŽ{n,g6y< rvҶ(-p|p<Yr1U ܹXX:7=Y| Z i}uv2Q5 Dd_ tXeޅJ"Y_ݖC %p%}@yu1 #cFAfKf,F!3 Fus_!_n@WTMz(?\# Li g<QU*uQLUiΔg(D*Wr!/.JrF~$gF`bFsyʌsnaKf'ZM5i3Љ<2O5jyD\7Dc_ɂ>dzZٗE8MqJhVb\ŔKZZY5% 2]$QKHk\q:e?ɑ1 CI@(  2 QF,}*b֖.V[ 2 tGԐe|^m4 UP:aIZGJצRj]qys++:t UR:J{}6NЗR>"vDX92tdlmsUWiuB1YΨ|0ljQ30 `yP yc9R** Y[%o \ @s%\|<:kޱY x ELLۗO\q5QƠ޾#%s\tR"mV~k$5+x 0Df9{19r-yvt FPq`9Z@UDc.8*q^o PHc{iiKq`WҾ\ar2@5qm_@VD`d4A p$cHmlgDb, =./2;B}y~m3+h[e鹤Yv!_z"jLmK.P:Cu[{CFZG]ԕY* K^qͭS孼J="ɣK۾G.i.k)FS+ҩgikFPFBψ ٛP;F MZ=uFtE[^X͛CzW g;!JxKJ**R]|%!B[x~Hp" "8օ4H~ъjq֡33bȍR8#wK0*nMψVhӛPkDp}ē4c}($ ?]*p@u:@BT[ǒᨬv:]qC-mE@齾54ߐt-H'ߊtM:vuw@ɤPJo7yq;%ӆ<((b!ҋX^$vW-\#ޖZKI]4 2\"]|͇I:uKZ)H5灜q;HV** ʇXrT%FrTDN|y Ģ ~=Q?3*ǥ1޽x#!Npja3aun`ψ+%Z#ud- Z]bpΠC&?dUiV*`aYQ$&EQ-@fYu]>Zũ1Tk W͠9cSBp>[d&bC(Fst#|r.wZsuiJDs𺈽 ?quPPa VL䨘fX_.V.YG8uQI6++9E[F$+"}* Fy"t?C0E1Mhz#qhks\>A縞B}'9\}l$omrrJzvpWä\.>|9Q 'v|=Q2)V)uy "jO|#3SbQ:'"Fq=󬚔x1EKRaLt\kZ+۟$>)C{ЖF nJm)pY޸%xK*9x=-+*II׻)H"~^^Ghe,l>Tɟ!o)̆'D}o0<0,D%: S+³GMR~{5ץ/޸ R)؄}Pl?7e~3x bt&@V)jȄů%1J^6 L"ĐKm=֚;x0&࿽|;(K *CTq4Ioؿ~zLVz"jPW;JdwI:5~V82{!\OfQY(Ewu>5p۸SUFvZv'/wA3TH.; s{kdFV!vZRofV)Xo`0/fV)5j25KZҭ1+8@+֔\ (rB!"/OygP.uwk̅Q/,O>\DMlza-2;Df˚Z?7;ޠoO<͒i 'MfըiZN?F `}/k^\\Ii 5/v2_+e'df,z;;\E;V*'\*ET&[/ %X7/l&)nAT7JkoA6,pt_mn@ߪݨB-Kgfv>EJ>|NQ(@Χ"|C흞_^=rnr|^8Nf7sK{[-P(Bͷ-ml#>`6gplץ}ۮԬ-t'z25+FLT{kMR7xcHo5< SEɜ$C,i*=G:Ϋ'}@G2ĢoHGJ,3s X¬ppc6͟|-Y{m za5\Zݓ(GOܿLqe+Ibn3%sWd9k\%g =8uwImgozuAkT|7߂ ,Xmurkrx%AWX7Xؖ~@jT(Ǧ9=6r)LGk#xv57r^]B*52zJy

ܝח31{!yzkùC\0V2u)C>G|Fe!2pdzJ{z gRLJukkZ'NIQN%t`< 6ac(0 5yw?U*QY+7$2v>y߻QKnpL'1muրsywB^)Rd|VGTyG~C֍wSp{;r{| Uys 3K[,dwҰ+/a=-/ي1Ty _@$dRY^u8';~ Z9w[.gZ "1aqkpR)3pgZ}3>xj-ACvjs>+MU~>Hۀ6*1dwod6( [ef{Ǘ4h[uDim^O!-!`ܶN{kYh+ ˜誆"#m~aHH.󂋽m`Y#@FN_!(&76Dޤ39S$uLcVҢT ڨGyQLR%skQz2J늢pT!9J:N!)Y*{K=hg4` WyI9+T^Jho"' YikcأBL k֝)T끫9(ʲ’UIPa$j:bEbÈwgoxnZ;(f4#Efn)_/"̦^iց Y.1YnCPY5*ȲT2ܛ|AiD,pOoS]q W .?׷OzN @"П~~yyU?|ݻ7J(TRCy~£TVs.2{kqUtO>ZٰNS9Sđuء͏=6ے꧛]l;Nꎤʼn];stÄ1ϔo-)0;m&k!@ȵ[N^ 6h'vej[zXoOR-o&VӓkFhT?[,Gu}nr-v~F*e 2O\ŵwFbݓA$` E'jqh!^@v8yId!U6,[F?b}Hupgs.%{Ȗ3Ÿc)<26i^j mu?dج?d[;kG.|" cଏDiSRU霴HU! ;HUW UWjflA^&fl xɽ61 ~jGaVN˲S=%uGX'Do.^=/zӪ2|AYZ_S*%~*!t{RT.P%A E=儝z{ JHr$!/&rKrS[rU`WYKhOPj —G0)'PoRԑouEɳTt NcF&zU{;.ŅꆾxyD_$;?+ž]03]#P|\ֆ$D)6Vṵ Y7QA:/ zZ= KӠQVjp4hJ9u(+@);Ì`m)Ke4B"PXBצ,cMHH Giv_jNF ,1pT.Ƙ*PY-E}@SdCB" !SU̬\d((ˀQy8&4ڢ*T?-+L0YpHRLam$ͩ%S@KiSՌPv7}U8>5+m{t}tZKA~I~Gk6mǟդ ާṚKݍ5t[t¾R_IpiGc՜P?ܳ1X&Ax.~xQtÏ/{x(A)5LMy2N(>9;yJsZѲԘ14,oJR@K˻N -D: 8\RxYOy-<39㝱goܔ1FIi:[J$N)q>ah.'4` k/Sd)tiڥcj#uC[Wbl?bZr|},;O_Ya}&ƾ8,V1ȿRD~u^I#-Uڽ&ca_603h;z6Sp-HLG} 0\X8ePҘ"7E!ja, fJf*(2UĐ@"BS۪TUh @ p2TJ <.Ldu^:kZyE2 gF&T(6dY3E8Y̳ܕMrrg.8!PYgbjDT+kůssFXPvC]e-K2EF+@Ֆ+HШv6 A~LJ,_!@3W EiՑo<~ Vm,syZvA`i AD^6iȢ,f"#`dެnz&D{_cHr-P c ^cPD t!ߑ%T(/[K,ʡ*Kyɞ8Z?AfYhtU!rҕpqnU eIcߪY] l&(ɦ~`l,ˀʔ:pv ak/EK-%M08ò[K++ZL݂}zךsEiH.jӬMLph-i-^Š3xi[Ѯi.\!C' < c,3Hkb,F<@iUtت~AdqjF5 YC!+-}l%O,)ҐUY-| YC!+͕!\BB|Xz| :r6V_ߜULLjM~[cMmOu՛z]]14L+6)vXڻڭfٶ`/| ,$,ɪ0{(uGC(d% a%D `+}U;]+9fg,dXȹd0dRZY OKK)&Rp#F0RK$0][ "4ͺ8EHT5:-Bm-U&BwPYߘͨpJAs(dp1RZe:{r'Uu̴,z|,l.fBFMNYR]}fIc .n6J1w%yK">av>(Ԣ\sK,H\zm(pq,\:Rm۝8NsZ/HH?Ww%\ aYZ=4;hD#:x'v(1[w%5N/18 ]ݜLրs Z;#NNgw]5^o;MQJ߅ȧ՗wGkJu`H4"qû揭~guS:k]sdAzF=80*`X9rA uF9HG1MpzxO+GЍi5p,Ll8L-Q,i۩ ѭvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005337220315153626477017722 0ustar rootrootMar 09 18:25:17 crc systemd[1]: Starting Kubernetes Kubelet... Mar 09 18:25:17 crc restorecon[4681]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:17 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 18:25:18 crc restorecon[4681]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 09 18:25:19 crc kubenswrapper[4750]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 18:25:19 crc kubenswrapper[4750]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 09 18:25:19 crc kubenswrapper[4750]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 18:25:19 crc kubenswrapper[4750]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 18:25:19 crc kubenswrapper[4750]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 09 18:25:19 crc kubenswrapper[4750]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.110601 4750 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117823 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117856 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117871 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117883 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117894 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117906 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117916 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117927 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117938 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117948 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117961 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117972 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117982 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.117992 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118007 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118024 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118036 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118047 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118058 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118070 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118081 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118098 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118111 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118121 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118132 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118159 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118170 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118180 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118190 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118200 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118210 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118222 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118232 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118242 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118252 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118262 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118272 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118282 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118292 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118302 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118314 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118325 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118336 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118346 4750 feature_gate.go:330] unrecognized feature gate: Example Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118360 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118374 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118386 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118398 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118409 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118419 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118429 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118439 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118450 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118461 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118474 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118485 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118518 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118533 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118545 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118561 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118571 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118582 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118592 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118603 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118613 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118623 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118664 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118675 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118686 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118696 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.118706 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.118916 4750 flags.go:64] FLAG: --address="0.0.0.0" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.118940 4750 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.118961 4750 flags.go:64] FLAG: --anonymous-auth="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.118975 4750 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.118989 4750 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119001 4750 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119015 4750 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119030 4750 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119042 4750 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119054 4750 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119066 4750 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119078 4750 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119091 4750 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119102 4750 flags.go:64] FLAG: --cgroup-root="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119113 4750 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119125 4750 flags.go:64] FLAG: --client-ca-file="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119137 4750 flags.go:64] FLAG: --cloud-config="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119150 4750 flags.go:64] FLAG: --cloud-provider="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.119162 4750 flags.go:64] FLAG: --cluster-dns="[]" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120125 4750 flags.go:64] FLAG: --cluster-domain="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120140 4750 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120154 4750 flags.go:64] FLAG: --config-dir="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120166 4750 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120181 4750 flags.go:64] FLAG: --container-log-max-files="5" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120199 4750 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120211 4750 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120223 4750 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120235 4750 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120248 4750 flags.go:64] FLAG: --contention-profiling="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120259 4750 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120270 4750 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120283 4750 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120294 4750 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120309 4750 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120321 4750 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120332 4750 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120344 4750 flags.go:64] FLAG: --enable-load-reader="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120355 4750 flags.go:64] FLAG: --enable-server="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120367 4750 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120384 4750 flags.go:64] FLAG: --event-burst="100" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120396 4750 flags.go:64] FLAG: --event-qps="50" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120408 4750 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120421 4750 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120432 4750 flags.go:64] FLAG: --eviction-hard="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120447 4750 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120458 4750 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120470 4750 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120483 4750 flags.go:64] FLAG: --eviction-soft="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120496 4750 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120509 4750 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120521 4750 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120533 4750 flags.go:64] FLAG: --experimental-mounter-path="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120545 4750 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120556 4750 flags.go:64] FLAG: --fail-swap-on="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120569 4750 flags.go:64] FLAG: --feature-gates="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120584 4750 flags.go:64] FLAG: --file-check-frequency="20s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120596 4750 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120607 4750 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120620 4750 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120667 4750 flags.go:64] FLAG: --healthz-port="10248" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120684 4750 flags.go:64] FLAG: --help="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120696 4750 flags.go:64] FLAG: --hostname-override="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120708 4750 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120719 4750 flags.go:64] FLAG: --http-check-frequency="20s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120731 4750 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120743 4750 flags.go:64] FLAG: --image-credential-provider-config="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120754 4750 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120766 4750 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120777 4750 flags.go:64] FLAG: --image-service-endpoint="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120788 4750 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120800 4750 flags.go:64] FLAG: --kube-api-burst="100" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120812 4750 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120824 4750 flags.go:64] FLAG: --kube-api-qps="50" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120836 4750 flags.go:64] FLAG: --kube-reserved="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120848 4750 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120860 4750 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120872 4750 flags.go:64] FLAG: --kubelet-cgroups="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120883 4750 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120895 4750 flags.go:64] FLAG: --lock-file="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120907 4750 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120919 4750 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120933 4750 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120952 4750 flags.go:64] FLAG: --log-json-split-stream="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120963 4750 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120975 4750 flags.go:64] FLAG: --log-text-split-stream="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120987 4750 flags.go:64] FLAG: --logging-format="text" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.120998 4750 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121011 4750 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121023 4750 flags.go:64] FLAG: --manifest-url="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121034 4750 flags.go:64] FLAG: --manifest-url-header="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121049 4750 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121062 4750 flags.go:64] FLAG: --max-open-files="1000000" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121076 4750 flags.go:64] FLAG: --max-pods="110" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121088 4750 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121100 4750 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121111 4750 flags.go:64] FLAG: --memory-manager-policy="None" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121126 4750 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121138 4750 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121149 4750 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121162 4750 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121190 4750 flags.go:64] FLAG: --node-status-max-images="50" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121202 4750 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121214 4750 flags.go:64] FLAG: --oom-score-adj="-999" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121225 4750 flags.go:64] FLAG: --pod-cidr="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121236 4750 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121253 4750 flags.go:64] FLAG: --pod-manifest-path="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121264 4750 flags.go:64] FLAG: --pod-max-pids="-1" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121276 4750 flags.go:64] FLAG: --pods-per-core="0" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121288 4750 flags.go:64] FLAG: --port="10250" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121300 4750 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121311 4750 flags.go:64] FLAG: --provider-id="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121324 4750 flags.go:64] FLAG: --qos-reserved="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121335 4750 flags.go:64] FLAG: --read-only-port="10255" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121347 4750 flags.go:64] FLAG: --register-node="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121360 4750 flags.go:64] FLAG: --register-schedulable="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121372 4750 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121391 4750 flags.go:64] FLAG: --registry-burst="10" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121402 4750 flags.go:64] FLAG: --registry-qps="5" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121414 4750 flags.go:64] FLAG: --reserved-cpus="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121425 4750 flags.go:64] FLAG: --reserved-memory="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121439 4750 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121451 4750 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121462 4750 flags.go:64] FLAG: --rotate-certificates="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121474 4750 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121485 4750 flags.go:64] FLAG: --runonce="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121497 4750 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121509 4750 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121521 4750 flags.go:64] FLAG: --seccomp-default="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121532 4750 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121552 4750 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121564 4750 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121576 4750 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121591 4750 flags.go:64] FLAG: --storage-driver-password="root" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121602 4750 flags.go:64] FLAG: --storage-driver-secure="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121614 4750 flags.go:64] FLAG: --storage-driver-table="stats" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121625 4750 flags.go:64] FLAG: --storage-driver-user="root" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121667 4750 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121681 4750 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121693 4750 flags.go:64] FLAG: --system-cgroups="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121704 4750 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121724 4750 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121736 4750 flags.go:64] FLAG: --tls-cert-file="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121748 4750 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121762 4750 flags.go:64] FLAG: --tls-min-version="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121773 4750 flags.go:64] FLAG: --tls-private-key-file="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121784 4750 flags.go:64] FLAG: --topology-manager-policy="none" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121797 4750 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121808 4750 flags.go:64] FLAG: --topology-manager-scope="container" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121821 4750 flags.go:64] FLAG: --v="2" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121836 4750 flags.go:64] FLAG: --version="false" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121851 4750 flags.go:64] FLAG: --vmodule="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121864 4750 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.121876 4750 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122146 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122161 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122173 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122184 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122194 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122210 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122222 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122233 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122249 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122259 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122270 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122280 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122290 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122300 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122310 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122322 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122336 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122349 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122360 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122370 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122381 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122391 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122402 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122412 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122422 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122433 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122444 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122454 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122464 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122474 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122484 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122494 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122504 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122514 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122524 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122534 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122545 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122556 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122567 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122577 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122591 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122601 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122611 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122621 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122662 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122673 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122683 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122693 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122704 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122714 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122725 4750 feature_gate.go:330] unrecognized feature gate: Example Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122737 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122748 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122758 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122768 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122782 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122796 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122812 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122824 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122834 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122844 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122855 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122865 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122876 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122886 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122896 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122906 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122917 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122927 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122937 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.122947 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.123884 4750 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.133941 4750 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.133972 4750 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134091 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134102 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134114 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134123 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134133 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134142 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134151 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134159 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134169 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134179 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134187 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134195 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134204 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134214 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134222 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134230 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134238 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134247 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134255 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134263 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134271 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134279 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134286 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134294 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134302 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134310 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134317 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134327 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134335 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134346 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134355 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134364 4750 feature_gate.go:330] unrecognized feature gate: Example Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134375 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134385 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134394 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134403 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134413 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134423 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134431 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134441 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134449 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134459 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134467 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134475 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134483 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134492 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134500 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134508 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134516 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134524 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134535 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134544 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134554 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134562 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134570 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134579 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134587 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134595 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134604 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134613 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134622 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134651 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134659 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134667 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134676 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134684 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134693 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134700 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134708 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134716 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.134724 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.134737 4750 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138238 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138282 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138292 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138302 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138314 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138331 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138340 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138348 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138356 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138365 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138373 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138381 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138389 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138403 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138412 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138420 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138428 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138436 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138444 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138453 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138461 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138469 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138477 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138485 4750 feature_gate.go:330] unrecognized feature gate: Example Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138494 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138502 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138515 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138523 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138531 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138539 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138547 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138555 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138563 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138571 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138582 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138593 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138614 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138623 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138654 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138668 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138679 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138689 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138698 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138707 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.138720 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139029 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139040 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139050 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139058 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139066 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139075 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139082 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139090 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139097 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139106 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139114 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139123 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139134 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139144 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139152 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139160 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139169 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139178 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139187 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139195 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139203 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139211 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139219 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139227 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139234 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.139242 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.139255 4750 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.139526 4750 server.go:940] "Client rotation is on, will bootstrap in background" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.144044 4750 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.148505 4750 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.148691 4750 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.151005 4750 server.go:997] "Starting client certificate rotation" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.151052 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.151600 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.180372 4750 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.183381 4750 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.184104 4750 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.206578 4750 log.go:25] "Validated CRI v1 runtime API" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.247555 4750 log.go:25] "Validated CRI v1 image API" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.250372 4750 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.255164 4750 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-09-18-20-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.255203 4750 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.274598 4750 manager.go:217] Machine: {Timestamp:2026-03-09 18:25:19.27225351 +0000 UTC m=+0.614725918 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:316dc856-891a-4700-afe9-6ff306b27388 BootID:836d638e-2eed-4bd3-92fd-26e5289ded19 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:20:7d:8b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:20:7d:8b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:33:c7:62 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ba:0c:45 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:62:84:81 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:4c:04:89 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d2:ac:10:94:92:0e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:2e:8f:83:e1:fc:b5 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.274829 4750 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.274957 4750 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.276013 4750 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.276170 4750 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.276197 4750 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.276420 4750 topology_manager.go:138] "Creating topology manager with none policy" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.276431 4750 container_manager_linux.go:303] "Creating device plugin manager" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.276906 4750 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.276935 4750 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.277619 4750 state_mem.go:36] "Initialized new in-memory state store" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.277739 4750 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.283332 4750 kubelet.go:418] "Attempting to sync node with API server" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.283372 4750 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.283417 4750 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.283435 4750 kubelet.go:324] "Adding apiserver pod source" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.283450 4750 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.287738 4750 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.289113 4750 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.290179 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.290259 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.290554 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.290855 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.293145 4750 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.294936 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.294979 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295000 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295021 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295047 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295063 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295079 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295104 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295122 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295138 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295187 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.295205 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.296488 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.297263 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.297338 4750 server.go:1280] "Started kubelet" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.298437 4750 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.298459 4750 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.299664 4750 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 09 18:25:19 crc systemd[1]: Started Kubernetes Kubelet. Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.302362 4750 server.go:460] "Adding debug handlers to kubelet server" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.303174 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.303216 4750 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.303690 4750 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.303725 4750 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.303699 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.303848 4750 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.308613 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.308720 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.308789 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="200ms" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.311341 4750 factory.go:55] Registering systemd factory Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.311400 4750 factory.go:221] Registration of the systemd container factory successfully Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.314987 4750 factory.go:153] Registering CRI-O factory Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.315048 4750 factory.go:221] Registration of the crio container factory successfully Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.315378 4750 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.315444 4750 factory.go:103] Registering Raw factory Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.315474 4750 manager.go:1196] Started watching for new ooms in manager Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.314926 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.93:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189b3f851fcf16d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,LastTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.316671 4750 manager.go:319] Starting recovery of all containers Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320129 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320260 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320357 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320435 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320507 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320587 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320720 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320808 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320889 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.320962 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321042 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321119 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321194 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321282 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321359 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321439 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321513 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321598 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321710 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321793 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.321889 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322015 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322129 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322238 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322351 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322446 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322548 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322663 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322748 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.322985 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.323094 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.323223 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.323321 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.323439 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.323602 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.323717 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.323925 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324034 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324270 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324397 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324503 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324612 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324747 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324865 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.324950 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325038 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325124 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325203 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325283 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325361 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325448 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325619 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.325734 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326672 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326719 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326735 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326749 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326765 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326784 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326798 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326814 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326828 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326849 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326865 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326883 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326901 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326916 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326931 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326946 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326965 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326982 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.326997 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327012 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327026 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327042 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327063 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327078 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327094 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327110 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327125 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327141 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327157 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327172 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327189 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327203 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327217 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327234 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327249 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327265 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327281 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327309 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327327 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327343 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327358 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327375 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327394 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327409 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327432 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327449 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327465 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327479 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327493 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327509 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327522 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327586 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.327611 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328428 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328476 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328502 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328528 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328554 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328575 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328597 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328662 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328703 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328732 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328763 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328795 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328815 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328837 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328871 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328890 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328909 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328932 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.328953 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331725 4750 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331803 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331830 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331862 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331925 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331945 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331965 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.331985 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332005 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332024 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332044 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332092 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332119 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332157 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332194 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332225 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332250 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332289 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332310 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332330 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332350 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332369 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332388 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332418 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332440 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332468 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332490 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332621 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332700 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332737 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332765 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332806 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332845 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332874 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332895 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332926 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332948 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.332968 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333018 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333040 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333062 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333084 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333105 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333126 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333155 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333176 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333211 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333237 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333276 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333308 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333346 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333373 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333399 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333428 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333472 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333497 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333528 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333551 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333571 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333590 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333686 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333710 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333731 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333751 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333772 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333808 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333828 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333852 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333874 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333903 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333942 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333972 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.333997 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334018 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334040 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334061 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334081 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334542 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334564 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334584 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334616 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334667 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334687 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334708 4750 reconstruct.go:97] "Volume reconstruction finished" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.334725 4750 reconciler.go:26] "Reconciler: start to sync state" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.349495 4750 manager.go:324] Recovery completed Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.364972 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.367075 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.367132 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.367148 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.368109 4750 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.368128 4750 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.368151 4750 state_mem.go:36] "Initialized new in-memory state store" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.369228 4750 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.371948 4750 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.372023 4750 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.372068 4750 kubelet.go:2335] "Starting kubelet main sync loop" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.372237 4750 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.373267 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.373408 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.383478 4750 policy_none.go:49] "None policy: Start" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.385672 4750 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.385725 4750 state_mem.go:35] "Initializing new in-memory state store" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.404782 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.447203 4750 manager.go:334] "Starting Device Plugin manager" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.447676 4750 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.447716 4750 server.go:79] "Starting device plugin registration server" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.448469 4750 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.448499 4750 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.448896 4750 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.449039 4750 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.449061 4750 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.459517 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.472394 4750 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.472547 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.473746 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.473781 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.473791 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.473902 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.474733 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.474759 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.474767 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.475368 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.475377 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.475434 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.475736 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.475531 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.476199 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.476241 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.476251 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.476337 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.476474 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.476531 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477324 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477368 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477381 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477393 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477404 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477400 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477538 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477378 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477710 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477771 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477810 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477866 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.477820 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478037 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478646 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478669 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478679 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478781 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478813 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478826 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478834 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.478855 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.479611 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.479671 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.479689 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.509742 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="400ms" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537337 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537376 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537394 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537415 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537437 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537454 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537536 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537572 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537605 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537669 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537718 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537735 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537753 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537774 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.537794 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.548801 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.550009 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.550129 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.550190 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.550259 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.550731 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.93:6443: connect: connection refused" node="crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.638975 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639060 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639111 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639155 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639197 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639240 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639282 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639325 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639367 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639408 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639452 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639531 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639585 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639704 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639708 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639731 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639759 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639819 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639857 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639841 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639912 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639906 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639982 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639890 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.640004 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.640016 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639889 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.639993 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.640109 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.640095 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.751272 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.753133 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.753197 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.753211 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.753247 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.753878 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.93:6443: connect: connection refused" node="crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.810103 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.822958 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.829442 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.846370 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: I0309 18:25:19.851678 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.867554 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c4815cea7376c0ae2a7bfa199bd7a6342d67e180c23aad8d0ba671b59724df64 WatchSource:0}: Error finding container c4815cea7376c0ae2a7bfa199bd7a6342d67e180c23aad8d0ba671b59724df64: Status 404 returned error can't find the container with id c4815cea7376c0ae2a7bfa199bd7a6342d67e180c23aad8d0ba671b59724df64 Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.869258 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ed2d53014728fed7a7ee8b4e26deddf2a04acc0d9442dfc9a89313cd0a9bca22 WatchSource:0}: Error finding container ed2d53014728fed7a7ee8b4e26deddf2a04acc0d9442dfc9a89313cd0a9bca22: Status 404 returned error can't find the container with id ed2d53014728fed7a7ee8b4e26deddf2a04acc0d9442dfc9a89313cd0a9bca22 Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.880078 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e8d6642951fb66d6b6f7a0d7cfcd378d6630a3876b6d5c6d0e1f80b12a587412 WatchSource:0}: Error finding container e8d6642951fb66d6b6f7a0d7cfcd378d6630a3876b6d5c6d0e1f80b12a587412: Status 404 returned error can't find the container with id e8d6642951fb66d6b6f7a0d7cfcd378d6630a3876b6d5c6d0e1f80b12a587412 Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.888329 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-3b96447cad3551a542eda961a0947b4b6e4f1809f859c7f81689488ebbffa1b3 WatchSource:0}: Error finding container 3b96447cad3551a542eda961a0947b4b6e4f1809f859c7f81689488ebbffa1b3: Status 404 returned error can't find the container with id 3b96447cad3551a542eda961a0947b4b6e4f1809f859c7f81689488ebbffa1b3 Mar 09 18:25:19 crc kubenswrapper[4750]: W0309 18:25:19.889287 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-c7692f20ac42ffc93c6d0b835db711370b17422157f30000b53f199a05b8342f WatchSource:0}: Error finding container c7692f20ac42ffc93c6d0b835db711370b17422157f30000b53f199a05b8342f: Status 404 returned error can't find the container with id c7692f20ac42ffc93c6d0b835db711370b17422157f30000b53f199a05b8342f Mar 09 18:25:19 crc kubenswrapper[4750]: E0309 18:25:19.911387 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="800ms" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.154597 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.156255 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.156326 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.156339 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.156368 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:20 crc kubenswrapper[4750]: E0309 18:25:20.156844 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.93:6443: connect: connection refused" node="crc" Mar 09 18:25:20 crc kubenswrapper[4750]: W0309 18:25:20.226345 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:20 crc kubenswrapper[4750]: E0309 18:25:20.226487 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:20 crc kubenswrapper[4750]: W0309 18:25:20.284582 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:20 crc kubenswrapper[4750]: E0309 18:25:20.284717 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.298687 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.377100 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c7692f20ac42ffc93c6d0b835db711370b17422157f30000b53f199a05b8342f"} Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.379254 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e8d6642951fb66d6b6f7a0d7cfcd378d6630a3876b6d5c6d0e1f80b12a587412"} Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.381305 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ed2d53014728fed7a7ee8b4e26deddf2a04acc0d9442dfc9a89313cd0a9bca22"} Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.383501 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c4815cea7376c0ae2a7bfa199bd7a6342d67e180c23aad8d0ba671b59724df64"} Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.385003 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3b96447cad3551a542eda961a0947b4b6e4f1809f859c7f81689488ebbffa1b3"} Mar 09 18:25:20 crc kubenswrapper[4750]: W0309 18:25:20.542585 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:20 crc kubenswrapper[4750]: E0309 18:25:20.542854 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:20 crc kubenswrapper[4750]: W0309 18:25:20.648700 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:20 crc kubenswrapper[4750]: E0309 18:25:20.648841 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:20 crc kubenswrapper[4750]: E0309 18:25:20.713130 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="1.6s" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.957407 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.961020 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.961081 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.961096 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:20 crc kubenswrapper[4750]: I0309 18:25:20.961131 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:20 crc kubenswrapper[4750]: E0309 18:25:20.961701 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.93:6443: connect: connection refused" node="crc" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.248421 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 18:25:21 crc kubenswrapper[4750]: E0309 18:25:21.250061 4750 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.298149 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.398689 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b485435fc91d1546f9ae0588d54051d12e8e12d9c241f685065b33d8ef9ae886"} Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.398870 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940"} Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.400770 4750 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26" exitCode=0 Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.400839 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26"} Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.400925 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.402125 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.402155 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.402169 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.403570 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3" exitCode=0 Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.403811 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3"} Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.404212 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.406775 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.406812 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.406832 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.407244 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119"} Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.407181 4750 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119" exitCode=0 Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.407657 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.409860 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.409915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.409934 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.410445 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.411232 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.411268 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.411282 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.411328 4750 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280" exitCode=0 Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.411367 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280"} Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.411418 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.412373 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.412410 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:21 crc kubenswrapper[4750]: I0309 18:25:21.412419 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.298221 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:22 crc kubenswrapper[4750]: E0309 18:25:22.314685 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="3.2s" Mar 09 18:25:22 crc kubenswrapper[4750]: W0309 18:25:22.388217 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:22 crc kubenswrapper[4750]: E0309 18:25:22.388329 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.419004 4750 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4" exitCode=0 Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.419139 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.419146 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.420216 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.420246 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.420256 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.423118 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.423179 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.423202 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.423216 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.424017 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.424044 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.424057 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.427506 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.427551 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.427647 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.428959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.429004 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.429018 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.429728 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.429878 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.430989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.431020 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.431030 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.433602 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.433803 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.433831 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.433850 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad"} Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.562268 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.564995 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.565036 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.565071 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:22 crc kubenswrapper[4750]: I0309 18:25:22.565106 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:22 crc kubenswrapper[4750]: E0309 18:25:22.565791 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.93:6443: connect: connection refused" node="crc" Mar 09 18:25:22 crc kubenswrapper[4750]: W0309 18:25:22.738071 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.93:6443: connect: connection refused Mar 09 18:25:22 crc kubenswrapper[4750]: E0309 18:25:22.738188 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.93:6443: connect: connection refused" logger="UnhandledError" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.440238 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"aa635a5fbc50946264136f5574df7d5d31f3ceb70460feb043e4fd89c575928b"} Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.440329 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.441525 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.441571 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.441583 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.442862 4750 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e" exitCode=0 Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.442921 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e"} Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.442952 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.443003 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.443026 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.443060 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.443026 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.444480 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.444512 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.444523 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.444938 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.444990 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.445018 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.445005 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.445029 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.445054 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.445068 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.445046 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:23 crc kubenswrapper[4750]: I0309 18:25:23.445180 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.451367 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"899a0632bb66ad4c1454d05ab958515d304030a7fa0ddd190d9a28133e4af8f7"} Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.451425 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7e617cc775c5b425a56aacf7fb7a049a8117366cee205d7289d896d04386608a"} Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.451441 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1577b46bd60cd09246da1ba3c6d76c6da8782969bbefe301ded194092da54f70"} Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.451452 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"171f973093d7d57ceae547565c16f8a12f0889ad5d0ad2b870320faad0e1cef4"} Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.451478 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.451680 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.451765 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.452281 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.452309 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.452319 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.453268 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.453299 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:24 crc kubenswrapper[4750]: I0309 18:25:24.453307 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.462338 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"531e2e851c9d42b4d3dc21665d1d36326a0c63d6cef342c9560c26590f8256b5"} Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.462550 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.464239 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.464285 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.464299 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.633254 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.766778 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.768780 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.768904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.768924 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.768972 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.792892 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.793103 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.793155 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.794872 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.794940 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:25 crc kubenswrapper[4750]: I0309 18:25:25.794966 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.181482 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.466707 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.466814 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.466851 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.468801 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.468883 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.468900 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.468895 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.469049 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.469075 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.648454 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.648780 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.650600 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.650722 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.650751 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:26 crc kubenswrapper[4750]: I0309 18:25:26.874091 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.374451 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.469755 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.469755 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.471359 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.471429 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.471455 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.471732 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.471787 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.471804 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.685433 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.685727 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.687059 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.687122 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:27 crc kubenswrapper[4750]: I0309 18:25:27.687136 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:29 crc kubenswrapper[4750]: E0309 18:25:29.459754 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:25:29 crc kubenswrapper[4750]: I0309 18:25:29.481783 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:29 crc kubenswrapper[4750]: I0309 18:25:29.481968 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:29 crc kubenswrapper[4750]: I0309 18:25:29.483690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:29 crc kubenswrapper[4750]: I0309 18:25:29.483808 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:29 crc kubenswrapper[4750]: I0309 18:25:29.483870 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:29 crc kubenswrapper[4750]: I0309 18:25:29.490374 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.481176 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.481509 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.483120 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.483216 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.483241 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.486297 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.686062 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:25:30 crc kubenswrapper[4750]: I0309 18:25:30.686193 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 09 18:25:31 crc kubenswrapper[4750]: I0309 18:25:31.484502 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:31 crc kubenswrapper[4750]: I0309 18:25:31.486210 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:31 crc kubenswrapper[4750]: I0309 18:25:31.486260 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:31 crc kubenswrapper[4750]: I0309 18:25:31.486278 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.122129 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.122544 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.487557 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.493110 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.493181 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.493203 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:32 crc kubenswrapper[4750]: W0309 18:25:32.896413 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.896546 4750 trace.go:236] Trace[1811771641]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Mar-2026 18:25:22.894) (total time: 10001ms): Mar 09 18:25:32 crc kubenswrapper[4750]: Trace[1811771641]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:25:32.896) Mar 09 18:25:32 crc kubenswrapper[4750]: Trace[1811771641]: [10.001557688s] [10.001557688s] END Mar 09 18:25:32 crc kubenswrapper[4750]: E0309 18:25:32.896588 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 09 18:25:32 crc kubenswrapper[4750]: W0309 18:25:32.901288 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 09 18:25:32 crc kubenswrapper[4750]: I0309 18:25:32.901422 4750 trace.go:236] Trace[921269406]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Mar-2026 18:25:22.900) (total time: 10001ms): Mar 09 18:25:32 crc kubenswrapper[4750]: Trace[921269406]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:25:32.901) Mar 09 18:25:32 crc kubenswrapper[4750]: Trace[921269406]: [10.001305858s] [10.001305858s] END Mar 09 18:25:32 crc kubenswrapper[4750]: E0309 18:25:32.901454 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 09 18:25:33 crc kubenswrapper[4750]: I0309 18:25:33.300135 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 09 18:25:33 crc kubenswrapper[4750]: E0309 18:25:33.783047 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b3f851fcf16d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,LastTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:25:33 crc kubenswrapper[4750]: W0309 18:25:33.783903 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z Mar 09 18:25:33 crc kubenswrapper[4750]: E0309 18:25:33.784022 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:33 crc kubenswrapper[4750]: W0309 18:25:33.784867 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z Mar 09 18:25:33 crc kubenswrapper[4750]: E0309 18:25:33.784964 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:33 crc kubenswrapper[4750]: I0309 18:25:33.785786 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 09 18:25:33 crc kubenswrapper[4750]: I0309 18:25:33.785876 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 09 18:25:33 crc kubenswrapper[4750]: E0309 18:25:33.789621 4750 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:33 crc kubenswrapper[4750]: E0309 18:25:33.791035 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 09 18:25:33 crc kubenswrapper[4750]: E0309 18:25:33.791999 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:33Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 18:25:33 crc kubenswrapper[4750]: I0309 18:25:33.794397 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 09 18:25:33 crc kubenswrapper[4750]: I0309 18:25:33.794463 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.301785 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:34Z is after 2026-02-23T05:33:13Z Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.495607 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.498234 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="aa635a5fbc50946264136f5574df7d5d31f3ceb70460feb043e4fd89c575928b" exitCode=255 Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.498292 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"aa635a5fbc50946264136f5574df7d5d31f3ceb70460feb043e4fd89c575928b"} Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.498470 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.499819 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.499872 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.499892 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:34 crc kubenswrapper[4750]: I0309 18:25:34.500722 4750 scope.go:117] "RemoveContainer" containerID="aa635a5fbc50946264136f5574df7d5d31f3ceb70460feb043e4fd89c575928b" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.259323 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.259690 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.263672 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.263743 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.263766 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.299489 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.301476 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:35Z is after 2026-02-23T05:33:13Z Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.505112 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.507304 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362"} Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.507434 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.507480 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.508567 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.508645 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.508658 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.508724 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.508775 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.508795 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.528126 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 09 18:25:35 crc kubenswrapper[4750]: I0309 18:25:35.799342 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:36 crc kubenswrapper[4750]: W0309 18:25:36.154473 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:36Z is after 2026-02-23T05:33:13Z Mar 09 18:25:36 crc kubenswrapper[4750]: E0309 18:25:36.154590 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:36Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.185972 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.300725 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:36Z is after 2026-02-23T05:33:13Z Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.513470 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.514388 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.517403 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362" exitCode=255 Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.517550 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.517554 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362"} Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.517653 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.517746 4750 scope.go:117] "RemoveContainer" containerID="aa635a5fbc50946264136f5574df7d5d31f3ceb70460feb043e4fd89c575928b" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.518803 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.518864 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.518890 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.519059 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.519118 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.519145 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:36 crc kubenswrapper[4750]: I0309 18:25:36.520045 4750 scope.go:117] "RemoveContainer" containerID="3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362" Mar 09 18:25:36 crc kubenswrapper[4750]: E0309 18:25:36.520485 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.302010 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:37Z is after 2026-02-23T05:33:13Z Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.374973 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.522409 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.525294 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.526658 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.526809 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.526915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:37 crc kubenswrapper[4750]: I0309 18:25:37.527937 4750 scope.go:117] "RemoveContainer" containerID="3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362" Mar 09 18:25:37 crc kubenswrapper[4750]: E0309 18:25:37.528372 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:25:38 crc kubenswrapper[4750]: I0309 18:25:38.303809 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:38Z is after 2026-02-23T05:33:13Z Mar 09 18:25:38 crc kubenswrapper[4750]: W0309 18:25:38.505142 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:38Z is after 2026-02-23T05:33:13Z Mar 09 18:25:38 crc kubenswrapper[4750]: E0309 18:25:38.505273 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:38Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:38 crc kubenswrapper[4750]: I0309 18:25:38.528434 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:38 crc kubenswrapper[4750]: I0309 18:25:38.530093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:38 crc kubenswrapper[4750]: I0309 18:25:38.530153 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:38 crc kubenswrapper[4750]: I0309 18:25:38.530173 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:38 crc kubenswrapper[4750]: I0309 18:25:38.531235 4750 scope.go:117] "RemoveContainer" containerID="3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362" Mar 09 18:25:38 crc kubenswrapper[4750]: E0309 18:25:38.531537 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:25:39 crc kubenswrapper[4750]: I0309 18:25:39.304013 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:39Z is after 2026-02-23T05:33:13Z Mar 09 18:25:39 crc kubenswrapper[4750]: E0309 18:25:39.460430 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.192402 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.194237 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.194285 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.194301 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.194336 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:40 crc kubenswrapper[4750]: E0309 18:25:40.196484 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:40Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 18:25:40 crc kubenswrapper[4750]: E0309 18:25:40.199228 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:40Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.302381 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:40Z is after 2026-02-23T05:33:13Z Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.687520 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:25:40 crc kubenswrapper[4750]: I0309 18:25:40.687733 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 18:25:41 crc kubenswrapper[4750]: I0309 18:25:41.302289 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:41Z is after 2026-02-23T05:33:13Z Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.121877 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.122848 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.124884 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.124957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.124979 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.126101 4750 scope.go:117] "RemoveContainer" containerID="3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362" Mar 09 18:25:42 crc kubenswrapper[4750]: E0309 18:25:42.126468 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.304938 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:42Z is after 2026-02-23T05:33:13Z Mar 09 18:25:42 crc kubenswrapper[4750]: I0309 18:25:42.511154 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 18:25:42 crc kubenswrapper[4750]: E0309 18:25:42.517427 4750 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:42Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:42 crc kubenswrapper[4750]: W0309 18:25:42.830295 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:42Z is after 2026-02-23T05:33:13Z Mar 09 18:25:42 crc kubenswrapper[4750]: E0309 18:25:42.830401 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:42Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:43 crc kubenswrapper[4750]: I0309 18:25:43.304071 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:43Z is after 2026-02-23T05:33:13Z Mar 09 18:25:43 crc kubenswrapper[4750]: E0309 18:25:43.790710 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:43Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b3f851fcf16d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,LastTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:25:44 crc kubenswrapper[4750]: W0309 18:25:44.061666 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:44Z is after 2026-02-23T05:33:13Z Mar 09 18:25:44 crc kubenswrapper[4750]: E0309 18:25:44.061769 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:44Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:44 crc kubenswrapper[4750]: I0309 18:25:44.302892 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:44Z is after 2026-02-23T05:33:13Z Mar 09 18:25:45 crc kubenswrapper[4750]: I0309 18:25:45.304151 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:45Z is after 2026-02-23T05:33:13Z Mar 09 18:25:45 crc kubenswrapper[4750]: W0309 18:25:45.511675 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:45Z is after 2026-02-23T05:33:13Z Mar 09 18:25:45 crc kubenswrapper[4750]: E0309 18:25:45.511833 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:45Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:45 crc kubenswrapper[4750]: W0309 18:25:45.970706 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:45Z is after 2026-02-23T05:33:13Z Mar 09 18:25:45 crc kubenswrapper[4750]: E0309 18:25:45.970809 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:45Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:46 crc kubenswrapper[4750]: I0309 18:25:46.304367 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:46Z is after 2026-02-23T05:33:13Z Mar 09 18:25:47 crc kubenswrapper[4750]: I0309 18:25:47.200309 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:47 crc kubenswrapper[4750]: E0309 18:25:47.201084 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:47Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 18:25:47 crc kubenswrapper[4750]: I0309 18:25:47.202335 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:47 crc kubenswrapper[4750]: I0309 18:25:47.202444 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:47 crc kubenswrapper[4750]: I0309 18:25:47.202491 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:47 crc kubenswrapper[4750]: I0309 18:25:47.202542 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:47 crc kubenswrapper[4750]: E0309 18:25:47.205543 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:47Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 18:25:47 crc kubenswrapper[4750]: I0309 18:25:47.302398 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:47Z is after 2026-02-23T05:33:13Z Mar 09 18:25:48 crc kubenswrapper[4750]: I0309 18:25:48.304083 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:48Z is after 2026-02-23T05:33:13Z Mar 09 18:25:49 crc kubenswrapper[4750]: I0309 18:25:49.304074 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:49Z is after 2026-02-23T05:33:13Z Mar 09 18:25:49 crc kubenswrapper[4750]: E0309 18:25:49.460827 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.303226 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:50Z is after 2026-02-23T05:33:13Z Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.687068 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.687207 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.687288 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.687487 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.689280 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.689365 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.689406 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.690747 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"b485435fc91d1546f9ae0588d54051d12e8e12d9c241f685065b33d8ef9ae886"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 09 18:25:50 crc kubenswrapper[4750]: I0309 18:25:50.691007 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://b485435fc91d1546f9ae0588d54051d12e8e12d9c241f685065b33d8ef9ae886" gracePeriod=30 Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.303092 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:51Z is after 2026-02-23T05:33:13Z Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.569026 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.569709 4750 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b485435fc91d1546f9ae0588d54051d12e8e12d9c241f685065b33d8ef9ae886" exitCode=255 Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.569779 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b485435fc91d1546f9ae0588d54051d12e8e12d9c241f685065b33d8ef9ae886"} Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.569826 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050"} Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.570033 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.571421 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.571480 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:51 crc kubenswrapper[4750]: I0309 18:25:51.571498 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:52 crc kubenswrapper[4750]: I0309 18:25:52.304925 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:52Z is after 2026-02-23T05:33:13Z Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.301559 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:53Z is after 2026-02-23T05:33:13Z Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.372742 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.374370 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.374460 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.374479 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.375336 4750 scope.go:117] "RemoveContainer" containerID="3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.579707 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.581987 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856"} Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.582157 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.583050 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.583076 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:53 crc kubenswrapper[4750]: I0309 18:25:53.583085 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:53 crc kubenswrapper[4750]: E0309 18:25:53.794486 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:53Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b3f851fcf16d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,LastTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.206439 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.208562 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.208612 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.208622 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.208667 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:25:54 crc kubenswrapper[4750]: E0309 18:25:54.209792 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:54Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 18:25:54 crc kubenswrapper[4750]: E0309 18:25:54.211832 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:54Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.303669 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:54Z is after 2026-02-23T05:33:13Z Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.588457 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.589501 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.592482 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856" exitCode=255 Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.592549 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856"} Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.592682 4750 scope.go:117] "RemoveContainer" containerID="3e472fc3c521e1fbb0f49b441538d170bfb062606d5a1244566c7842f9c49362" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.592857 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.593967 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.593999 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.594008 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:54 crc kubenswrapper[4750]: I0309 18:25:54.594537 4750 scope.go:117] "RemoveContainer" containerID="f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856" Mar 09 18:25:54 crc kubenswrapper[4750]: E0309 18:25:54.595322 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:25:55 crc kubenswrapper[4750]: I0309 18:25:55.304384 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:55Z is after 2026-02-23T05:33:13Z Mar 09 18:25:55 crc kubenswrapper[4750]: I0309 18:25:55.598222 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 18:25:56 crc kubenswrapper[4750]: I0309 18:25:56.304403 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:56Z is after 2026-02-23T05:33:13Z Mar 09 18:25:56 crc kubenswrapper[4750]: I0309 18:25:56.649037 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:56 crc kubenswrapper[4750]: I0309 18:25:56.649306 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:56 crc kubenswrapper[4750]: I0309 18:25:56.651224 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:56 crc kubenswrapper[4750]: I0309 18:25:56.651275 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:56 crc kubenswrapper[4750]: I0309 18:25:56.651292 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.302998 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:57Z is after 2026-02-23T05:33:13Z Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.375240 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.375477 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.376975 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.377023 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.377042 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.377742 4750 scope.go:117] "RemoveContainer" containerID="f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856" Mar 09 18:25:57 crc kubenswrapper[4750]: E0309 18:25:57.377988 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.685906 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.686213 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.688261 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.688323 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:25:57 crc kubenswrapper[4750]: I0309 18:25:57.688337 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:25:58 crc kubenswrapper[4750]: I0309 18:25:58.302515 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:58Z is after 2026-02-23T05:33:13Z Mar 09 18:25:59 crc kubenswrapper[4750]: I0309 18:25:59.302353 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:59Z is after 2026-02-23T05:33:13Z Mar 09 18:25:59 crc kubenswrapper[4750]: E0309 18:25:59.460971 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:25:59 crc kubenswrapper[4750]: I0309 18:25:59.651215 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 18:25:59 crc kubenswrapper[4750]: E0309 18:25:59.657757 4750 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:25:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:25:59 crc kubenswrapper[4750]: E0309 18:25:59.658970 4750 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 09 18:26:00 crc kubenswrapper[4750]: W0309 18:26:00.156493 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:00Z is after 2026-02-23T05:33:13Z Mar 09 18:26:00 crc kubenswrapper[4750]: E0309 18:26:00.156998 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:26:00 crc kubenswrapper[4750]: I0309 18:26:00.301566 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:00Z is after 2026-02-23T05:33:13Z Mar 09 18:26:00 crc kubenswrapper[4750]: W0309 18:26:00.618033 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:00Z is after 2026-02-23T05:33:13Z Mar 09 18:26:00 crc kubenswrapper[4750]: E0309 18:26:00.618132 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:26:00 crc kubenswrapper[4750]: I0309 18:26:00.686124 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:26:00 crc kubenswrapper[4750]: I0309 18:26:00.686201 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 18:26:01 crc kubenswrapper[4750]: I0309 18:26:01.212229 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:01 crc kubenswrapper[4750]: I0309 18:26:01.214344 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:01 crc kubenswrapper[4750]: I0309 18:26:01.214407 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:01 crc kubenswrapper[4750]: I0309 18:26:01.214425 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:01 crc kubenswrapper[4750]: I0309 18:26:01.214463 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:26:01 crc kubenswrapper[4750]: E0309 18:26:01.214822 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:01Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 18:26:01 crc kubenswrapper[4750]: E0309 18:26:01.218658 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:01Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 18:26:01 crc kubenswrapper[4750]: I0309 18:26:01.301270 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:01Z is after 2026-02-23T05:33:13Z Mar 09 18:26:02 crc kubenswrapper[4750]: I0309 18:26:02.121785 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:26:02 crc kubenswrapper[4750]: I0309 18:26:02.122021 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:02 crc kubenswrapper[4750]: I0309 18:26:02.123384 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:02 crc kubenswrapper[4750]: I0309 18:26:02.123439 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:02 crc kubenswrapper[4750]: I0309 18:26:02.123452 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:02 crc kubenswrapper[4750]: I0309 18:26:02.124132 4750 scope.go:117] "RemoveContainer" containerID="f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856" Mar 09 18:26:02 crc kubenswrapper[4750]: E0309 18:26:02.124324 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:26:02 crc kubenswrapper[4750]: I0309 18:26:02.302306 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:02Z is after 2026-02-23T05:33:13Z Mar 09 18:26:03 crc kubenswrapper[4750]: I0309 18:26:03.300766 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:03Z is after 2026-02-23T05:33:13Z Mar 09 18:26:03 crc kubenswrapper[4750]: E0309 18:26:03.801624 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:03Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b3f851fcf16d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,LastTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:04 crc kubenswrapper[4750]: I0309 18:26:04.302390 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:04Z is after 2026-02-23T05:33:13Z Mar 09 18:26:04 crc kubenswrapper[4750]: W0309 18:26:04.488191 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:04Z is after 2026-02-23T05:33:13Z Mar 09 18:26:04 crc kubenswrapper[4750]: E0309 18:26:04.488682 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:26:05 crc kubenswrapper[4750]: I0309 18:26:05.301859 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:05Z is after 2026-02-23T05:33:13Z Mar 09 18:26:06 crc kubenswrapper[4750]: I0309 18:26:06.304549 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:06Z is after 2026-02-23T05:33:13Z Mar 09 18:26:07 crc kubenswrapper[4750]: W0309 18:26:07.100477 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:07Z is after 2026-02-23T05:33:13Z Mar 09 18:26:07 crc kubenswrapper[4750]: E0309 18:26:07.100677 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:07Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 18:26:07 crc kubenswrapper[4750]: I0309 18:26:07.301160 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:07Z is after 2026-02-23T05:33:13Z Mar 09 18:26:08 crc kubenswrapper[4750]: I0309 18:26:08.218849 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:08 crc kubenswrapper[4750]: E0309 18:26:08.218851 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:08Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 18:26:08 crc kubenswrapper[4750]: I0309 18:26:08.220486 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:08 crc kubenswrapper[4750]: I0309 18:26:08.220534 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:08 crc kubenswrapper[4750]: I0309 18:26:08.220546 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:08 crc kubenswrapper[4750]: I0309 18:26:08.220583 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:26:08 crc kubenswrapper[4750]: E0309 18:26:08.223494 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:08Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 18:26:08 crc kubenswrapper[4750]: I0309 18:26:08.303224 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:08Z is after 2026-02-23T05:33:13Z Mar 09 18:26:09 crc kubenswrapper[4750]: I0309 18:26:09.301552 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:09Z is after 2026-02-23T05:33:13Z Mar 09 18:26:09 crc kubenswrapper[4750]: E0309 18:26:09.461101 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:26:10 crc kubenswrapper[4750]: I0309 18:26:10.302785 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:10Z is after 2026-02-23T05:33:13Z Mar 09 18:26:10 crc kubenswrapper[4750]: I0309 18:26:10.686247 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:26:10 crc kubenswrapper[4750]: I0309 18:26:10.686340 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 18:26:11 crc kubenswrapper[4750]: I0309 18:26:11.304026 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:11Z is after 2026-02-23T05:33:13Z Mar 09 18:26:12 crc kubenswrapper[4750]: I0309 18:26:12.206356 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 18:26:12 crc kubenswrapper[4750]: I0309 18:26:12.206578 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:12 crc kubenswrapper[4750]: I0309 18:26:12.207959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:12 crc kubenswrapper[4750]: I0309 18:26:12.207993 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:12 crc kubenswrapper[4750]: I0309 18:26:12.208009 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:12 crc kubenswrapper[4750]: I0309 18:26:12.303502 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:12Z is after 2026-02-23T05:33:13Z Mar 09 18:26:13 crc kubenswrapper[4750]: I0309 18:26:13.303053 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:26:13Z is after 2026-02-23T05:33:13Z Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.808280 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f851fcf16d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,LastTimestamp:2026-03-09 18:25:19.297287889 +0000 UTC m=+0.639760317,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.816986 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.821454 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.825262 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f92ee7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,LastTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.829121 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8528f6f981 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.450896769 +0000 UTC m=+0.793369197,LastTimestamp:2026-03-09 18:25:19.450896769 +0000 UTC m=+0.793369197,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.833539 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f86b8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.473770003 +0000 UTC m=+0.816242401,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.837895 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f8fdd6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.473788414 +0000 UTC m=+0.816260802,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.841159 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f92ee7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f92ee7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,LastTimestamp:2026-03-09 18:25:19.473796394 +0000 UTC m=+0.816268792,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.846135 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f86b8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.474754389 +0000 UTC m=+0.817226787,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.851161 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f8fdd6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.474764149 +0000 UTC m=+0.817236547,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.855195 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f92ee7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f92ee7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,LastTimestamp:2026-03-09 18:25:19.474771549 +0000 UTC m=+0.817243947,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.862023 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f86b8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.476227737 +0000 UTC m=+0.818700135,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.867566 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f8fdd6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.476247998 +0000 UTC m=+0.818720396,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.872850 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f92ee7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f92ee7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,LastTimestamp:2026-03-09 18:25:19.476256338 +0000 UTC m=+0.818728736,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.877595 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f86b8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.477350567 +0000 UTC m=+0.819823045,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.882355 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f8fdd6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.477374347 +0000 UTC m=+0.819846745,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.888413 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f86b8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.477386578 +0000 UTC m=+0.819858976,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.893920 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f86b8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.477394318 +0000 UTC m=+0.819866716,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.901207 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f8fdd6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.477401208 +0000 UTC m=+0.819873606,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.905239 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f92ee7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f92ee7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,LastTimestamp:2026-03-09 18:25:19.477412308 +0000 UTC m=+0.819884706,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.909464 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f8fdd6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.477522841 +0000 UTC m=+0.819995249,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.913351 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f92ee7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f92ee7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,LastTimestamp:2026-03-09 18:25:19.477548732 +0000 UTC m=+0.820021120,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.917599 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f92ee7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f92ee7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367155431 +0000 UTC m=+0.709627839,LastTimestamp:2026-03-09 18:25:19.477557012 +0000 UTC m=+0.820029410,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.921763 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f86b8b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f86b8b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.367105419 +0000 UTC m=+0.709577827,LastTimestamp:2026-03-09 18:25:19.477799878 +0000 UTC m=+0.820272306,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.925446 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b3f8523f8fdd6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b3f8523f8fdd6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.36714287 +0000 UTC m=+0.709615278,LastTimestamp:2026-03-09 18:25:19.477859329 +0000 UTC m=+0.820331767,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.930172 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f8542769626 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.878690342 +0000 UTC m=+1.221162730,LastTimestamp:2026-03-09 18:25:19.878690342 +0000 UTC m=+1.221162730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.934242 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8542795ce9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.878872297 +0000 UTC m=+1.221344705,LastTimestamp:2026-03-09 18:25:19.878872297 +0000 UTC m=+1.221344705,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.938580 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b3f8542ad28ac openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.882266796 +0000 UTC m=+1.224739214,LastTimestamp:2026-03-09 18:25:19.882266796 +0000 UTC m=+1.224739214,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.942177 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f8543899247 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.896711751 +0000 UTC m=+1.239184149,LastTimestamp:2026-03-09 18:25:19.896711751 +0000 UTC m=+1.239184149,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.945355 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8543bad67b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:19.899940475 +0000 UTC m=+1.242412873,LastTimestamp:2026-03-09 18:25:19.899940475 +0000 UTC m=+1.242412873,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.948710 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8567c6e516 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.504710422 +0000 UTC m=+1.847182870,LastTimestamp:2026-03-09 18:25:20.504710422 +0000 UTC m=+1.847182870,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.951542 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f8567cc602c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.505069612 +0000 UTC m=+1.847542050,LastTimestamp:2026-03-09 18:25:20.505069612 +0000 UTC m=+1.847542050,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.955864 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f8567d5ed9f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.505695647 +0000 UTC m=+1.848168085,LastTimestamp:2026-03-09 18:25:20.505695647 +0000 UTC m=+1.848168085,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.959471 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f85686adc3d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.515456061 +0000 UTC m=+1.857928459,LastTimestamp:2026-03-09 18:25:20.515456061 +0000 UTC m=+1.857928459,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.963616 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b3f856889df64 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.517488484 +0000 UTC m=+1.859960912,LastTimestamp:2026-03-09 18:25:20.517488484 +0000 UTC m=+1.859960912,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.968191 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f8568b67a8f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.520411791 +0000 UTC m=+1.862884219,LastTimestamp:2026-03-09 18:25:20.520411791 +0000 UTC m=+1.862884219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.972164 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8568c3724a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.521261642 +0000 UTC m=+1.863734080,LastTimestamp:2026-03-09 18:25:20.521261642 +0000 UTC m=+1.863734080,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.978310 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8568e64e50 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.523546192 +0000 UTC m=+1.866018630,LastTimestamp:2026-03-09 18:25:20.523546192 +0000 UTC m=+1.866018630,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.982842 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f8569822ba9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.533760937 +0000 UTC m=+1.876233345,LastTimestamp:2026-03-09 18:25:20.533760937 +0000 UTC m=+1.876233345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.986952 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8569d54b47 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.539208519 +0000 UTC m=+1.881680927,LastTimestamp:2026-03-09 18:25:20.539208519 +0000 UTC m=+1.881680927,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.991477 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b3f856a51c219 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.547365401 +0000 UTC m=+1.889837829,LastTimestamp:2026-03-09 18:25:20.547365401 +0000 UTC m=+1.889837829,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:13 crc kubenswrapper[4750]: E0309 18:26:13.995406 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8590983ce9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.189518569 +0000 UTC m=+2.531990977,LastTimestamp:2026-03-09 18:25:21.189518569 +0000 UTC m=+2.531990977,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.000060 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f85914b0efb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.201237755 +0000 UTC m=+2.543710153,LastTimestamp:2026-03-09 18:25:21.201237755 +0000 UTC m=+2.543710153,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.003543 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f85915deaff openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.202473727 +0000 UTC m=+2.544946125,LastTimestamp:2026-03-09 18:25:21.202473727 +0000 UTC m=+2.544946125,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.008128 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b3f859d6138d3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.404016851 +0000 UTC m=+2.746489259,LastTimestamp:2026-03-09 18:25:21.404016851 +0000 UTC m=+2.746489259,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.012483 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f859dc082b6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.410261686 +0000 UTC m=+2.752734084,LastTimestamp:2026-03-09 18:25:21.410261686 +0000 UTC m=+2.752734084,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.016154 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f859dd042a6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.411293862 +0000 UTC m=+2.753766260,LastTimestamp:2026-03-09 18:25:21.411293862 +0000 UTC m=+2.753766260,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.020170 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f859e2a618c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.417200012 +0000 UTC m=+2.759672440,LastTimestamp:2026-03-09 18:25:21.417200012 +0000 UTC m=+2.759672440,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.024704 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f859ea2c527 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.425089831 +0000 UTC m=+2.767562239,LastTimestamp:2026-03-09 18:25:21.425089831 +0000 UTC m=+2.767562239,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.029032 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f85a0127fc2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.449189314 +0000 UTC m=+2.791661712,LastTimestamp:2026-03-09 18:25:21.449189314 +0000 UTC m=+2.791661712,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.033219 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f85a02b0cd7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.450798295 +0000 UTC m=+2.793270743,LastTimestamp:2026-03-09 18:25:21.450798295 +0000 UTC m=+2.793270743,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.036876 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f85ab72e32a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.640055594 +0000 UTC m=+2.982527992,LastTimestamp:2026-03-09 18:25:21.640055594 +0000 UTC m=+2.982527992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.040540 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b3f85ab993427 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.642566695 +0000 UTC m=+2.985039093,LastTimestamp:2026-03-09 18:25:21.642566695 +0000 UTC m=+2.985039093,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.045344 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f85ab9d3ddf openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.642831327 +0000 UTC m=+2.985303725,LastTimestamp:2026-03-09 18:25:21.642831327 +0000 UTC m=+2.985303725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.049614 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85ab9dec37 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.642875959 +0000 UTC m=+2.985348357,LastTimestamp:2026-03-09 18:25:21.642875959 +0000 UTC m=+2.985348357,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.053951 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85ab9eac55 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.642925141 +0000 UTC m=+2.985397539,LastTimestamp:2026-03-09 18:25:21.642925141 +0000 UTC m=+2.985397539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.057679 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f85ac317ee9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.652547305 +0000 UTC m=+2.995019713,LastTimestamp:2026-03-09 18:25:21.652547305 +0000 UTC m=+2.995019713,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.061852 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85ac6664b9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.656014009 +0000 UTC m=+2.998486397,LastTimestamp:2026-03-09 18:25:21.656014009 +0000 UTC m=+2.998486397,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.066130 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85ac790860 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.657235552 +0000 UTC m=+2.999707950,LastTimestamp:2026-03-09 18:25:21.657235552 +0000 UTC m=+2.999707950,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.070189 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85aca812fb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.660318459 +0000 UTC m=+3.002790857,LastTimestamp:2026-03-09 18:25:21.660318459 +0000 UTC m=+3.002790857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.073473 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85acb48626 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.661134374 +0000 UTC m=+3.003606772,LastTimestamp:2026-03-09 18:25:21.661134374 +0000 UTC m=+3.003606772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.076709 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b3f85acd464a3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.663222947 +0000 UTC m=+3.005695345,LastTimestamp:2026-03-09 18:25:21.663222947 +0000 UTC m=+3.005695345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.080051 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85b7364fa5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.837412261 +0000 UTC m=+3.179884659,LastTimestamp:2026-03-09 18:25:21.837412261 +0000 UTC m=+3.179884659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.083441 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85b8069a63 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.851062883 +0000 UTC m=+3.193535281,LastTimestamp:2026-03-09 18:25:21.851062883 +0000 UTC m=+3.193535281,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.086617 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85b81c8147 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.852498247 +0000 UTC m=+3.194970655,LastTimestamp:2026-03-09 18:25:21.852498247 +0000 UTC m=+3.194970655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.090340 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85b849811f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.855447327 +0000 UTC m=+3.197919725,LastTimestamp:2026-03-09 18:25:21.855447327 +0000 UTC m=+3.197919725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.094433 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85b9b7ac05 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.879444485 +0000 UTC m=+3.221916883,LastTimestamp:2026-03-09 18:25:21.879444485 +0000 UTC m=+3.221916883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.098086 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85b9d00e5e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.881042526 +0000 UTC m=+3.223514924,LastTimestamp:2026-03-09 18:25:21.881042526 +0000 UTC m=+3.223514924,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.102125 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85c45a9d3d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.057895229 +0000 UTC m=+3.400367617,LastTimestamp:2026-03-09 18:25:22.057895229 +0000 UTC m=+3.400367617,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.106564 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85c54946f8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.073536248 +0000 UTC m=+3.416008646,LastTimestamp:2026-03-09 18:25:22.073536248 +0000 UTC m=+3.416008646,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.110000 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b3f85c5a6fb7d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.079677309 +0000 UTC m=+3.422149707,LastTimestamp:2026-03-09 18:25:22.079677309 +0000 UTC m=+3.422149707,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.113952 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85c5ebc8f6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.084186358 +0000 UTC m=+3.426658756,LastTimestamp:2026-03-09 18:25:22.084186358 +0000 UTC m=+3.426658756,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.117645 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85c5fcba39 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.085296697 +0000 UTC m=+3.427769085,LastTimestamp:2026-03-09 18:25:22.085296697 +0000 UTC m=+3.427769085,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.121122 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f85caceb8ad openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.166167725 +0000 UTC m=+3.508640123,LastTimestamp:2026-03-09 18:25:22.166167725 +0000 UTC m=+3.508640123,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.126327 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85cface334 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.247836468 +0000 UTC m=+3.590308876,LastTimestamp:2026-03-09 18:25:22.247836468 +0000 UTC m=+3.590308876,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.131365 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85d06aa94e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.260273486 +0000 UTC m=+3.602745894,LastTimestamp:2026-03-09 18:25:22.260273486 +0000 UTC m=+3.602745894,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.137256 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85d081d07b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.261790843 +0000 UTC m=+3.604263241,LastTimestamp:2026-03-09 18:25:22.261790843 +0000 UTC m=+3.604263241,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.156692 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f85da173c1f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.422578207 +0000 UTC m=+3.765050605,LastTimestamp:2026-03-09 18:25:22.422578207 +0000 UTC m=+3.765050605,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.161826 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85db968535 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.447697205 +0000 UTC m=+3.790169623,LastTimestamp:2026-03-09 18:25:22.447697205 +0000 UTC m=+3.790169623,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.168233 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85dc5c3d31 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.460654897 +0000 UTC m=+3.803127305,LastTimestamp:2026-03-09 18:25:22.460654897 +0000 UTC m=+3.803127305,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.172159 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f85e84b03ef openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.660852719 +0000 UTC m=+4.003325117,LastTimestamp:2026-03-09 18:25:22.660852719 +0000 UTC m=+4.003325117,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.176090 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f85ea3d63e3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.693514211 +0000 UTC m=+4.035986619,LastTimestamp:2026-03-09 18:25:22.693514211 +0000 UTC m=+4.035986619,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.180351 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f861729bfa5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:23.447201701 +0000 UTC m=+4.789674109,LastTimestamp:2026-03-09 18:25:23.447201701 +0000 UTC m=+4.789674109,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.183906 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f862624606f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:23.698507887 +0000 UTC m=+5.040980295,LastTimestamp:2026-03-09 18:25:23.698507887 +0000 UTC m=+5.040980295,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.187330 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8626de8c64 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:23.710708836 +0000 UTC m=+5.053181244,LastTimestamp:2026-03-09 18:25:23.710708836 +0000 UTC m=+5.053181244,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.190556 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8626f0831c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:23.711886108 +0000 UTC m=+5.054358516,LastTimestamp:2026-03-09 18:25:23.711886108 +0000 UTC m=+5.054358516,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.193729 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8635463fed openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:23.952386029 +0000 UTC m=+5.294858477,LastTimestamp:2026-03-09 18:25:23.952386029 +0000 UTC m=+5.294858477,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.196920 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8635fc939d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:23.964335005 +0000 UTC m=+5.306807413,LastTimestamp:2026-03-09 18:25:23.964335005 +0000 UTC m=+5.306807413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.200091 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8636159b01 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:23.965975297 +0000 UTC m=+5.308447725,LastTimestamp:2026-03-09 18:25:23.965975297 +0000 UTC m=+5.308447725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.203329 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f864329a27a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.185391738 +0000 UTC m=+5.527864136,LastTimestamp:2026-03-09 18:25:24.185391738 +0000 UTC m=+5.527864136,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.206986 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8643e8d2bf openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.197921471 +0000 UTC m=+5.540393879,LastTimestamp:2026-03-09 18:25:24.197921471 +0000 UTC m=+5.540393879,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.210540 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f8644034245 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.199653957 +0000 UTC m=+5.542126405,LastTimestamp:2026-03-09 18:25:24.199653957 +0000 UTC m=+5.542126405,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.213670 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f864fabdf5c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.395253596 +0000 UTC m=+5.737725994,LastTimestamp:2026-03-09 18:25:24.395253596 +0000 UTC m=+5.737725994,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.217131 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f865066a581 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.407494017 +0000 UTC m=+5.749966415,LastTimestamp:2026-03-09 18:25:24.407494017 +0000 UTC m=+5.749966415,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.220315 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f865077465e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.408583774 +0000 UTC m=+5.751056172,LastTimestamp:2026-03-09 18:25:24.408583774 +0000 UTC m=+5.751056172,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.223941 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f865cd3af6e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.615966574 +0000 UTC m=+5.958438972,LastTimestamp:2026-03-09 18:25:24.615966574 +0000 UTC m=+5.958438972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.227237 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b3f865dadeac7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:24.630268615 +0000 UTC m=+5.972741053,LastTimestamp:2026-03-09 18:25:24.630268615 +0000 UTC m=+5.972741053,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.232769 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-controller-manager-crc.189b3f87c6a35e8d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 09 18:26:14 crc kubenswrapper[4750]: body: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:30.686152333 +0000 UTC m=+12.028624771,LastTimestamp:2026-03-09 18:25:30.686152333 +0000 UTC m=+12.028624771,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.236377 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f87c6a7e7d5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:30.686449621 +0000 UTC m=+12.028922049,LastTimestamp:2026-03-09 18:25:30.686449621 +0000 UTC m=+12.028922049,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.239799 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-apiserver-crc.189b3f881c4083b1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 09 18:26:14 crc kubenswrapper[4750]: body: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:32.122514353 +0000 UTC m=+13.464986771,LastTimestamp:2026-03-09 18:25:32.122514353 +0000 UTC m=+13.464986771,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.243068 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f881c4423f5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:32.122751989 +0000 UTC m=+13.465224427,LastTimestamp:2026-03-09 18:25:32.122751989 +0000 UTC m=+13.465224427,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.247360 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-apiserver-crc.189b3f887f64dd63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 09 18:26:14 crc kubenswrapper[4750]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 09 18:26:14 crc kubenswrapper[4750]: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:33.785840995 +0000 UTC m=+15.128313433,LastTimestamp:2026-03-09 18:25:33.785840995 +0000 UTC m=+15.128313433,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.250668 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f887f65f5e7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:33.785912807 +0000 UTC m=+15.128385245,LastTimestamp:2026-03-09 18:25:33.785912807 +0000 UTC m=+15.128385245,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.254547 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189b3f887f64dd63\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-apiserver-crc.189b3f887f64dd63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 09 18:26:14 crc kubenswrapper[4750]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 09 18:26:14 crc kubenswrapper[4750]: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:33.785840995 +0000 UTC m=+15.128313433,LastTimestamp:2026-03-09 18:25:33.794445025 +0000 UTC m=+15.136917423,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.258325 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189b3f887f65f5e7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f887f65f5e7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:33.785912807 +0000 UTC m=+15.128385245,LastTimestamp:2026-03-09 18:25:33.794502397 +0000 UTC m=+15.136974795,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.262696 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189b3f85d081d07b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b3f85d081d07b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:22.261790843 +0000 UTC m=+3.604263241,LastTimestamp:2026-03-09 18:25:34.502042835 +0000 UTC m=+15.844515233,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.268042 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-controller-manager-crc.189b3f8a1ac5935f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 18:26:14 crc kubenswrapper[4750]: body: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:40.687614815 +0000 UTC m=+22.030087253,LastTimestamp:2026-03-09 18:25:40.687614815 +0000 UTC m=+22.030087253,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.271256 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8a1ac84000 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:40.68779008 +0000 UTC m=+22.030262518,LastTimestamp:2026-03-09 18:25:40.68779008 +0000 UTC m=+22.030262518,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.276689 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f8a1ac5935f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-controller-manager-crc.189b3f8a1ac5935f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 18:26:14 crc kubenswrapper[4750]: body: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:40.687614815 +0000 UTC m=+22.030087253,LastTimestamp:2026-03-09 18:25:50.687164936 +0000 UTC m=+32.029637334,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.280408 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f8a1ac84000\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8a1ac84000 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:40.68779008 +0000 UTC m=+22.030262518,LastTimestamp:2026-03-09 18:25:50.687246388 +0000 UTC m=+32.029718806,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.284603 4750 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8c6f04ed1d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:50.690987293 +0000 UTC m=+32.033459691,LastTimestamp:2026-03-09 18:25:50.690987293 +0000 UTC m=+32.033459691,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.289927 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f8568e64e50\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8568e64e50 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:20.523546192 +0000 UTC m=+1.866018630,LastTimestamp:2026-03-09 18:25:50.809403616 +0000 UTC m=+32.151876024,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.294509 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f8590983ce9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8590983ce9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.189518569 +0000 UTC m=+2.531990977,LastTimestamp:2026-03-09 18:25:51.001978169 +0000 UTC m=+32.344450577,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.298703 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f85914b0efb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f85914b0efb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:21.201237755 +0000 UTC m=+2.543710153,LastTimestamp:2026-03-09 18:25:51.011212146 +0000 UTC m=+32.353684544,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.304990 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f8a1ac5935f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-controller-manager-crc.189b3f8a1ac5935f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 18:26:14 crc kubenswrapper[4750]: body: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:40.687614815 +0000 UTC m=+22.030087253,LastTimestamp:2026-03-09 18:26:00.68617969 +0000 UTC m=+42.028652088,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:14 crc kubenswrapper[4750]: I0309 18:26:14.305311 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.309216 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f8a1ac84000\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b3f8a1ac84000 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:40.68779008 +0000 UTC m=+22.030262518,LastTimestamp:2026-03-09 18:26:00.686235402 +0000 UTC m=+42.028707800,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:26:14 crc kubenswrapper[4750]: E0309 18:26:14.315753 4750 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b3f8a1ac5935f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 18:26:14 crc kubenswrapper[4750]: &Event{ObjectMeta:{kube-controller-manager-crc.189b3f8a1ac5935f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 18:26:14 crc kubenswrapper[4750]: body: Mar 09 18:26:14 crc kubenswrapper[4750]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:25:40.687614815 +0000 UTC m=+22.030087253,LastTimestamp:2026-03-09 18:26:10.686312007 +0000 UTC m=+52.028784415,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 18:26:14 crc kubenswrapper[4750]: > Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.243760 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.245701 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.245745 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.245756 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.245791 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:26:15 crc kubenswrapper[4750]: E0309 18:26:15.250689 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 09 18:26:15 crc kubenswrapper[4750]: E0309 18:26:15.253467 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.299694 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.373403 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.375669 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.375730 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.375744 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.376543 4750 scope.go:117] "RemoveContainer" containerID="f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.664511 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.667060 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f"} Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.667253 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.668261 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.668325 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:15 crc kubenswrapper[4750]: I0309 18:26:15.668348 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.302545 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.673588 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.674223 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.676498 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" exitCode=255 Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.676602 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f"} Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.676781 4750 scope.go:117] "RemoveContainer" containerID="f0a5cff49b5fe03829c7153e1f3130200b8234ac34c087c78835195a2f5bb856" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.677081 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.678552 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.678714 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.678825 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:16 crc kubenswrapper[4750]: I0309 18:26:16.679487 4750 scope.go:117] "RemoveContainer" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" Mar 09 18:26:16 crc kubenswrapper[4750]: E0309 18:26:16.679755 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.304676 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.374782 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.682857 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.685370 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.686296 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.686325 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.686336 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:17 crc kubenswrapper[4750]: I0309 18:26:17.686958 4750 scope.go:117] "RemoveContainer" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" Mar 09 18:26:17 crc kubenswrapper[4750]: E0309 18:26:17.687129 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:26:18 crc kubenswrapper[4750]: I0309 18:26:18.304553 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:19 crc kubenswrapper[4750]: I0309 18:26:19.303748 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:19 crc kubenswrapper[4750]: E0309 18:26:19.461258 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.303168 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.686342 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.686484 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.686583 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.686888 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.688877 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.688918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.688932 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.689529 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 09 18:26:20 crc kubenswrapper[4750]: I0309 18:26:20.689652 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050" gracePeriod=30 Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.304458 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.706787 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.712071 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.713130 4750 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050" exitCode=255 Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.713203 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050"} Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.713252 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69"} Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.713282 4750 scope.go:117] "RemoveContainer" containerID="b485435fc91d1546f9ae0588d54051d12e8e12d9c241f685065b33d8ef9ae886" Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.713537 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.716094 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.716141 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:21 crc kubenswrapper[4750]: I0309 18:26:21.716159 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.122400 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.122789 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.124719 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.124769 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.124785 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.125593 4750 scope.go:117] "RemoveContainer" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" Mar 09 18:26:22 crc kubenswrapper[4750]: E0309 18:26:22.125853 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.251138 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.253427 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.253485 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.253504 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.253542 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:26:22 crc kubenswrapper[4750]: E0309 18:26:22.261702 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 09 18:26:22 crc kubenswrapper[4750]: E0309 18:26:22.261963 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.307271 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:22 crc kubenswrapper[4750]: I0309 18:26:22.719649 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 18:26:23 crc kubenswrapper[4750]: I0309 18:26:23.304244 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:24 crc kubenswrapper[4750]: I0309 18:26:24.302983 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:25 crc kubenswrapper[4750]: I0309 18:26:25.304772 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:26 crc kubenswrapper[4750]: I0309 18:26:26.302870 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:26 crc kubenswrapper[4750]: I0309 18:26:26.648847 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:26:26 crc kubenswrapper[4750]: I0309 18:26:26.649752 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:26 crc kubenswrapper[4750]: I0309 18:26:26.651260 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:26 crc kubenswrapper[4750]: I0309 18:26:26.651397 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:26 crc kubenswrapper[4750]: I0309 18:26:26.651418 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.302384 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.686209 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.686480 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.688203 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.688254 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.688352 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.692406 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.735367 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.740782 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.740843 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:27 crc kubenswrapper[4750]: I0309 18:26:27.740857 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:28 crc kubenswrapper[4750]: I0309 18:26:28.301967 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:29 crc kubenswrapper[4750]: I0309 18:26:29.262799 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:29 crc kubenswrapper[4750]: I0309 18:26:29.263925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:29 crc kubenswrapper[4750]: I0309 18:26:29.263956 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:29 crc kubenswrapper[4750]: I0309 18:26:29.263965 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:29 crc kubenswrapper[4750]: I0309 18:26:29.263989 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:26:29 crc kubenswrapper[4750]: E0309 18:26:29.267054 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 09 18:26:29 crc kubenswrapper[4750]: E0309 18:26:29.267126 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 09 18:26:29 crc kubenswrapper[4750]: I0309 18:26:29.299337 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:29 crc kubenswrapper[4750]: E0309 18:26:29.462017 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:26:30 crc kubenswrapper[4750]: I0309 18:26:30.303264 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:31 crc kubenswrapper[4750]: I0309 18:26:31.302528 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:31 crc kubenswrapper[4750]: W0309 18:26:31.391263 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 09 18:26:31 crc kubenswrapper[4750]: E0309 18:26:31.391320 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 09 18:26:31 crc kubenswrapper[4750]: I0309 18:26:31.661201 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 18:26:31 crc kubenswrapper[4750]: I0309 18:26:31.675055 4750 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 09 18:26:31 crc kubenswrapper[4750]: W0309 18:26:31.755318 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 09 18:26:31 crc kubenswrapper[4750]: E0309 18:26:31.755367 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 09 18:26:32 crc kubenswrapper[4750]: I0309 18:26:32.304023 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:33 crc kubenswrapper[4750]: I0309 18:26:33.302176 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.304803 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.373586 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.376151 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.376234 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.376257 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.431548 4750 csr.go:261] certificate signing request csr-5xm68 is approved, waiting to be issued Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.442242 4750 csr.go:257] certificate signing request csr-5xm68 is issued Mar 09 18:26:34 crc kubenswrapper[4750]: I0309 18:26:34.532076 4750 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 09 18:26:35 crc kubenswrapper[4750]: I0309 18:26:35.151356 4750 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 09 18:26:35 crc kubenswrapper[4750]: I0309 18:26:35.443739 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-03 11:42:21.023184036 +0000 UTC Mar 09 18:26:35 crc kubenswrapper[4750]: I0309 18:26:35.443834 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6449h15m45.579353131s for next certificate rotation Mar 09 18:26:35 crc kubenswrapper[4750]: I0309 18:26:35.943181 4750 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.268271 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.271069 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.271159 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.271179 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.271400 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.280373 4750 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.281135 4750 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.281170 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.285555 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.285595 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.285607 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.285653 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.285669 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:36Z","lastTransitionTime":"2026-03-09T18:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.299232 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.304502 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.304555 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.304569 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.304591 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.304608 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:36Z","lastTransitionTime":"2026-03-09T18:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.325180 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.331952 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.332004 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.332022 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.332044 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.332062 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:36Z","lastTransitionTime":"2026-03-09T18:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.348947 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.354748 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.355039 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.355342 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.355520 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.355761 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:36Z","lastTransitionTime":"2026-03-09T18:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.373529 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.374056 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.374139 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.475272 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.575794 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.653752 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.653961 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.655436 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.655483 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:36 crc kubenswrapper[4750]: I0309 18:26:36.655496 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.677016 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.777916 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.878324 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:36 crc kubenswrapper[4750]: E0309 18:26:36.979437 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.080401 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.180539 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.281349 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: I0309 18:26:37.373523 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:37 crc kubenswrapper[4750]: I0309 18:26:37.375327 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:37 crc kubenswrapper[4750]: I0309 18:26:37.375388 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:37 crc kubenswrapper[4750]: I0309 18:26:37.375399 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:37 crc kubenswrapper[4750]: I0309 18:26:37.376303 4750 scope.go:117] "RemoveContainer" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.376605 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.381448 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.481894 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.583065 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.683502 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.784148 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.884742 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:37 crc kubenswrapper[4750]: E0309 18:26:37.985966 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.087026 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.187423 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.287945 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.388823 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.489040 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.589242 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.690132 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.791038 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.891940 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:38 crc kubenswrapper[4750]: E0309 18:26:38.992762 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.092910 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.194037 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.294253 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.394869 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.462554 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.496302 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.597711 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.698303 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.798870 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:39 crc kubenswrapper[4750]: E0309 18:26:39.899847 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.000721 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.101525 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.201698 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.302734 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.403671 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.504773 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.605857 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.706032 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.806283 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:40 crc kubenswrapper[4750]: E0309 18:26:40.907048 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.007789 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.108761 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.208922 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.309245 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.409737 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.509844 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.610987 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.711955 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.812963 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:41 crc kubenswrapper[4750]: E0309 18:26:41.913615 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.014361 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.115085 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.215590 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.316076 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.416741 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.517215 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.618121 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.718441 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.819174 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:42 crc kubenswrapper[4750]: E0309 18:26:42.920304 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.021259 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.122354 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.222688 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.323312 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.424015 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.524186 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.625205 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.725686 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.826086 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: E0309 18:26:43.926718 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:43 crc kubenswrapper[4750]: I0309 18:26:43.972816 4750 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.027298 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.127929 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.228650 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.329223 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.429759 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.530590 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.631793 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.732252 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.833413 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:44 crc kubenswrapper[4750]: E0309 18:26:44.934499 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.035278 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.135499 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.236023 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.336149 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.436848 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.538294 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.638453 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.739614 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.840388 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:45 crc kubenswrapper[4750]: E0309 18:26:45.941549 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.042620 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.143502 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.243908 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.344531 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.447722 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.451919 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.457550 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.457925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.458031 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.458121 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.458200 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:46Z","lastTransitionTime":"2026-03-09T18:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.468186 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.472934 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.472979 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.472993 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.473016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.473031 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:46Z","lastTransitionTime":"2026-03-09T18:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.482901 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.488223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.488435 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.488563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.488699 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.488815 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:46Z","lastTransitionTime":"2026-03-09T18:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.498047 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.502358 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.502721 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.502862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.503002 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:46 crc kubenswrapper[4750]: I0309 18:26:46.503134 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:46Z","lastTransitionTime":"2026-03-09T18:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.516893 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.517581 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.548357 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.649553 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.750762 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.851756 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:46 crc kubenswrapper[4750]: E0309 18:26:46.952662 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.053811 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.154968 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.255997 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.357086 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.457520 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.558668 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.659070 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.759751 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.860510 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:47 crc kubenswrapper[4750]: E0309 18:26:47.961161 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.062357 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.162891 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.263404 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.364076 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.464716 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.565139 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.666609 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.767237 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.868898 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:48 crc kubenswrapper[4750]: E0309 18:26:48.969414 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.070586 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.171564 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.273306 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.374095 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.463586 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.474841 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.575235 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.676332 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.777076 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.877587 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:49 crc kubenswrapper[4750]: E0309 18:26:49.978796 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.079456 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.180569 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.281137 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.381560 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.481757 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.582396 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.682955 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.784122 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.884695 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:50 crc kubenswrapper[4750]: E0309 18:26:50.985227 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.086109 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.186288 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.287527 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.388338 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.488666 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.588925 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.689295 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.789578 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.890359 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:51 crc kubenswrapper[4750]: E0309 18:26:51.991576 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.092741 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.193251 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.294360 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: I0309 18:26:52.373508 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:26:52 crc kubenswrapper[4750]: I0309 18:26:52.374868 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:52 crc kubenswrapper[4750]: I0309 18:26:52.374973 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:52 crc kubenswrapper[4750]: I0309 18:26:52.374992 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:52 crc kubenswrapper[4750]: I0309 18:26:52.375692 4750 scope.go:117] "RemoveContainer" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.375923 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.395049 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.495725 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.596041 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.697220 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.798148 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:52 crc kubenswrapper[4750]: E0309 18:26:52.899194 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:52.999926 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.100687 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.201807 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.302200 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.403145 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.503708 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.604180 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.704341 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.805049 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:53 crc kubenswrapper[4750]: E0309 18:26:53.905891 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.006809 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.107820 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.208202 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.308763 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.409301 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.509759 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.610695 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.711900 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.812086 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:54 crc kubenswrapper[4750]: E0309 18:26:54.913821 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.014917 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.115289 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.215823 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.316804 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.417866 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.518603 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.619722 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.720133 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.820473 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:55 crc kubenswrapper[4750]: E0309 18:26:55.920624 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.021604 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.122558 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.223527 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.324603 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.425755 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.526187 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.552384 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.557752 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.557815 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.557833 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.557859 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.557877 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:56Z","lastTransitionTime":"2026-03-09T18:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.570751 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.574319 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.574351 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.574362 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.574379 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.574389 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:56Z","lastTransitionTime":"2026-03-09T18:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.584256 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.587421 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.587440 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.587448 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.587459 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.587466 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:56Z","lastTransitionTime":"2026-03-09T18:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.602385 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.606879 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.606902 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.606910 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.606925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:26:56 crc kubenswrapper[4750]: I0309 18:26:56.606935 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:26:56Z","lastTransitionTime":"2026-03-09T18:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.621731 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.621901 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.627218 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.728611 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.829505 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:56 crc kubenswrapper[4750]: E0309 18:26:56.930571 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.031701 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.132298 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.232465 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.332927 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.434010 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.535067 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.635443 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.736200 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.836862 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:57 crc kubenswrapper[4750]: E0309 18:26:57.937938 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.038443 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.138985 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.240068 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.340921 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.442045 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.542606 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.642986 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.743858 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.844907 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:58 crc kubenswrapper[4750]: E0309 18:26:58.945049 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.046214 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.146333 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.247247 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.347706 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.448834 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.464112 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.549045 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.649892 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.750913 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.852135 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:26:59 crc kubenswrapper[4750]: E0309 18:26:59.953109 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.054182 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.155151 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.255504 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.355918 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.456465 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.557441 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.658168 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.758663 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.858840 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:00 crc kubenswrapper[4750]: E0309 18:27:00.959234 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.059409 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.160603 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.261511 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.362216 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.463011 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.563956 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.664477 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.764939 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.865313 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:01 crc kubenswrapper[4750]: E0309 18:27:01.965430 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.066382 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.167131 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.267994 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.368697 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.469752 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.570866 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.671588 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.772159 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:02 crc kubenswrapper[4750]: E0309 18:27:02.873260 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:02.974358 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.075228 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.175438 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.276036 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: I0309 18:27:03.373361 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:27:03 crc kubenswrapper[4750]: I0309 18:27:03.375173 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:03 crc kubenswrapper[4750]: I0309 18:27:03.375236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:03 crc kubenswrapper[4750]: I0309 18:27:03.375286 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.376710 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.477530 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.578201 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.678728 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.779810 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.880873 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:03 crc kubenswrapper[4750]: E0309 18:27:03.981714 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.082344 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.182735 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.283579 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.384354 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.485307 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.586199 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.686400 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.787418 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.888313 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:04 crc kubenswrapper[4750]: E0309 18:27:04.988887 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.089927 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.191716 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.292275 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.372790 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.374088 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.374121 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.374133 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.374767 4750 scope.go:117] "RemoveContainer" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.392782 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.493716 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.594711 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.695771 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.796703 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.857125 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.860585 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba"} Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.860831 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.862563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.862682 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:05 crc kubenswrapper[4750]: I0309 18:27:05.862711 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.897170 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:05 crc kubenswrapper[4750]: E0309 18:27:05.997845 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.098206 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.198542 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.299523 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.400975 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.501732 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.601934 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.702323 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.778327 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.784029 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.784080 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.784098 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.784122 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.784141 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:06Z","lastTransitionTime":"2026-03-09T18:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.801557 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.806942 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.807042 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.807054 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.807081 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.807095 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:06Z","lastTransitionTime":"2026-03-09T18:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.824544 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.829262 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.829335 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.829358 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.829383 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.829404 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:06Z","lastTransitionTime":"2026-03-09T18:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.845760 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.855735 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.855777 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.855788 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.855804 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.855820 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:06Z","lastTransitionTime":"2026-03-09T18:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.866792 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.868099 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.870186 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.870395 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.870430 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.871537 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" exitCode=255 Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.871598 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba"} Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.871685 4750 scope.go:117] "RemoveContainer" containerID="294c65976644953ba3e88d9a4f7561579defee994fa77dfe846f355f28e24c3f" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.871914 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.873343 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.873404 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.873429 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:06 crc kubenswrapper[4750]: I0309 18:27:06.874436 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.874768 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:27:06 crc kubenswrapper[4750]: E0309 18:27:06.970846 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.071766 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.172590 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.273332 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.373816 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: I0309 18:27:07.375041 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.475084 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.576216 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.676943 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.777109 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: I0309 18:27:07.876381 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.877233 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:07 crc kubenswrapper[4750]: I0309 18:27:07.879122 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:27:07 crc kubenswrapper[4750]: I0309 18:27:07.880162 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:07 crc kubenswrapper[4750]: I0309 18:27:07.880203 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:07 crc kubenswrapper[4750]: I0309 18:27:07.880213 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:07 crc kubenswrapper[4750]: I0309 18:27:07.880885 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.881051 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:27:07 crc kubenswrapper[4750]: E0309 18:27:07.977590 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.078815 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.179040 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.280159 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.380270 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.481321 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.582522 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.683552 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.783774 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.884202 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:08 crc kubenswrapper[4750]: E0309 18:27:08.985297 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.086074 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.186786 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.287962 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.388602 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.465088 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.489005 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.589898 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.690795 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.791710 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.892361 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:09 crc kubenswrapper[4750]: E0309 18:27:09.993165 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.094063 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.195121 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.296322 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.397260 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.498097 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.598240 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.699411 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.800608 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:10 crc kubenswrapper[4750]: E0309 18:27:10.901180 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.001971 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.102858 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.203794 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.304894 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.405320 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.505886 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.606913 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.707471 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.808596 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:11 crc kubenswrapper[4750]: E0309 18:27:11.909041 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.009843 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.110683 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: I0309 18:27:12.121998 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:27:12 crc kubenswrapper[4750]: I0309 18:27:12.122293 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 18:27:12 crc kubenswrapper[4750]: I0309 18:27:12.124024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:12 crc kubenswrapper[4750]: I0309 18:27:12.124082 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:12 crc kubenswrapper[4750]: I0309 18:27:12.124093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:12 crc kubenswrapper[4750]: I0309 18:27:12.124959 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.125197 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.211317 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.311604 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.412372 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.512524 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.613606 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.714086 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.814955 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:12 crc kubenswrapper[4750]: E0309 18:27:12.915498 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.015783 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.116739 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.217887 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.318980 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.419956 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.520097 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.620688 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.721646 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.823112 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:13 crc kubenswrapper[4750]: E0309 18:27:13.923807 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.024427 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.124823 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.226002 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.326838 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.427650 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.527850 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.628669 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.729578 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.830505 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:14 crc kubenswrapper[4750]: E0309 18:27:14.931525 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.032431 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.133511 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.234317 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.335244 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.436390 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.537361 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.638021 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.739082 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.840103 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:15 crc kubenswrapper[4750]: E0309 18:27:15.940959 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.041751 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.142860 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.243722 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.344843 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.445317 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.546226 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.647045 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.747720 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.848750 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:16 crc kubenswrapper[4750]: E0309 18:27:16.949582 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.050764 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.078063 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.084090 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.084148 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.084167 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.084193 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.084211 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:17Z","lastTransitionTime":"2026-03-09T18:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.100863 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.106329 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.106440 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.106494 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.106523 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.106576 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:17Z","lastTransitionTime":"2026-03-09T18:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.125784 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.131118 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.131191 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.131211 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.131237 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.131256 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:17Z","lastTransitionTime":"2026-03-09T18:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.143877 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.149048 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.149101 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.149122 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.149146 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:17 crc kubenswrapper[4750]: I0309 18:27:17.149163 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:17Z","lastTransitionTime":"2026-03-09T18:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.160072 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.160247 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.160277 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.261192 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.362006 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.462424 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.563230 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.664148 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.765290 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.866184 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:17 crc kubenswrapper[4750]: E0309 18:27:17.966328 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.067320 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.167839 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.268707 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.369876 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.471115 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.572197 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.672882 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.773516 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.874541 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:18 crc kubenswrapper[4750]: E0309 18:27:18.974987 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:19 crc kubenswrapper[4750]: E0309 18:27:19.075617 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:19 crc kubenswrapper[4750]: E0309 18:27:19.176195 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:19 crc kubenswrapper[4750]: E0309 18:27:19.277031 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 18:27:19 crc kubenswrapper[4750]: E0309 18:27:19.378156 4750 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 09 18:27:19 crc kubenswrapper[4750]: E0309 18:27:19.465554 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 18:27:19 crc kubenswrapper[4750]: E0309 18:27:19.476444 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.181003 4750 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.350577 4750 apiserver.go:52] "Watching apiserver" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.356775 4750 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.357210 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-vf7m4","openshift-multus/multus-vj6kf","openshift-network-operator/iptables-alerter-4ln5h","openshift-machine-config-operator/machine-config-daemon-pqlpj","openshift-multus/network-metrics-daemon-wl7w6","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-lfwr6","openshift-image-registry/node-ca-4c5gf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r","openshift-multus/multus-additional-cni-plugins-ckgw2","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.357754 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.357793 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.357834 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.357925 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.357928 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.357977 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.358213 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.358279 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.358274 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.358439 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.359026 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.359086 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.359255 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.359478 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.359610 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.359858 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.360810 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.360559 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.360696 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.369344 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.369702 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.369966 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.370496 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.370610 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.370889 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.370959 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371089 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371150 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371223 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371269 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371434 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371490 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371549 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371613 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371681 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.371954 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372051 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372113 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372199 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372310 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372419 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372481 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372530 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372418 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.372747 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.375028 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.375196 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.375459 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.375754 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.375882 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.375988 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.379112 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.379287 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.379616 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.382208 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.405895 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.415205 4750 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.426124 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.457532 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460055 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460140 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460166 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460187 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460208 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460227 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460247 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.460604 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.461026 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.461348 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.461776 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.462303 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.462503 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463831 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463863 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463887 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463911 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463929 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463950 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463967 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463984 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464005 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464027 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464187 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464221 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464240 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464265 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464289 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464306 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464324 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464344 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464367 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464392 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464416 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.462810 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.462840 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.463773 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464156 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464425 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464439 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464890 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464920 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464950 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.464976 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465005 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465029 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465057 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465080 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465105 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465134 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465156 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465179 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465207 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465233 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465263 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465291 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465321 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465349 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465374 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465400 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465426 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465452 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465479 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465504 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465534 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465563 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465589 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465613 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465661 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465688 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465714 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465738 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465762 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466030 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466253 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466287 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466313 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466341 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466363 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466398 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466424 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466452 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466480 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466508 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466538 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466566 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466592 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466621 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466670 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466695 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466713 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466732 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466765 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466785 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466805 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466824 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466847 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466868 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466885 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466903 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466921 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466940 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466961 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466978 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466996 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467014 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467033 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467051 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467355 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467418 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467440 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467511 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467533 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467578 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467599 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467658 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467679 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467753 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469866 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469923 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470031 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470085 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470111 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470136 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470160 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470185 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470209 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470230 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470249 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470265 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470285 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470304 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470321 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470340 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470360 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470382 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470402 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470421 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470442 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470465 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470483 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470500 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470518 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470535 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470552 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470569 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470588 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470606 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470640 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470657 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470675 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470704 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470722 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470780 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470806 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470829 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470971 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470996 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471016 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471035 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471055 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471082 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471199 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471230 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471251 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471269 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471289 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471313 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471337 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471361 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471383 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471407 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471438 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471463 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471483 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471505 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471532 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471558 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471586 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471610 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471653 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471671 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471696 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471717 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471743 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471778 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471811 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471847 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471876 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471902 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471928 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471954 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471977 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471997 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472023 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472043 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472065 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472089 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472112 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472130 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472148 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472169 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472195 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472219 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472246 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472325 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7abc7c60-6863-4f05-b6d3-09bacb20a405-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472317 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472360 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472384 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472405 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v5n4\" (UniqueName: \"kubernetes.io/projected/80b0e746-0802-4112-b77b-4db85dc85a9c-kube-api-access-7v5n4\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472427 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472451 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fa652827-92a6-4544-b1b6-cba227f8e9f4-cni-binary-copy\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472479 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-ovn\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472644 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssdqs\" (UniqueName: \"kubernetes.io/projected/fa652827-92a6-4544-b1b6-cba227f8e9f4-kube-api-access-ssdqs\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472668 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-proxy-tls\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472687 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-script-lib\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472706 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-os-release\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472731 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472768 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-cni-multus\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472793 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-systemd-units\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473088 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473112 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-netns\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473138 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvbbj\" (UniqueName: \"kubernetes.io/projected/7abc7c60-6863-4f05-b6d3-09bacb20a405-kube-api-access-lvbbj\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473163 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473182 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/feb53af0-677a-4160-a671-f183573c68da-cni-binary-copy\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473212 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ec6607b-f406-49cd-84db-c83837e9735b-host\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473239 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-config\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473265 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-os-release\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473295 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473351 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-cnibin\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473383 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7abc7c60-6863-4f05-b6d3-09bacb20a405-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473414 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ec6607b-f406-49cd-84db-c83837e9735b-serviceca\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473444 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-daemon-config\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473488 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473517 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-etc-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473546 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sw7v\" (UniqueName: \"kubernetes.io/projected/cc616aab-babf-4203-823a-6d5e9d5edcd3-kube-api-access-2sw7v\") pod \"node-resolver-vf7m4\" (UID: \"cc616aab-babf-4203-823a-6d5e9d5edcd3\") " pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473613 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473681 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kn78\" (UniqueName: \"kubernetes.io/projected/feb53af0-677a-4160-a671-f183573c68da-kube-api-access-2kn78\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473703 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-hostroot\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473732 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473782 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-kubelet\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473802 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-netd\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465287 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465312 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465506 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465539 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474001 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466004 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466192 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466262 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466470 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466718 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466684 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466991 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.466999 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467125 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.467919 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.468300 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.468324 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.468114 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.468155 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.468558 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.468865 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469003 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469034 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469105 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469447 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469500 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469549 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.469601 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.470949 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471098 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474216 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471273 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471696 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471796 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471875 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472012 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.471918 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472132 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472158 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472190 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473822 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474591 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474689 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-cni-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474750 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95qv2\" (UniqueName: \"kubernetes.io/projected/6a32a985-7c9a-4737-86db-d46edf0d0565-kube-api-access-95qv2\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472381 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474789 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-rootfs\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472349 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474818 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-system-cni-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472444 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472604 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472716 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472742 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472918 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472961 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473057 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.472997 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473221 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473305 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473339 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473404 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473442 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473568 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473720 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473755 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.473811 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.465751 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474442 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474945 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474645 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474626 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474986 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474822 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474840 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.474871 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475120 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475015 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-node-log\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475504 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475529 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-log-socket\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475556 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475530 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475678 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475932 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.476000 4750 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.476176 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.476240 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.476615 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.476679 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.477012 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.477046 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.477336 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.477593 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.477610 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.477907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.477947 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.478117 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.478388 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.478700 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.478949 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.478945 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.479621 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.479975 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.480163 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.481598 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.481607 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.481675 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.482223 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.482218 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.482255 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.482644 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.483297 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.483808 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.483887 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.484134 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475580 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-conf-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.475578 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.484266 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.484752 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.485029 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.485191 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.485533 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.485715 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.485746 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.485998 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.486090 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.486129 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.486135 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.486654 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.486778 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.487055 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.487072 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.487146 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.487226 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.487065 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.487893 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.487965 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.488005 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.488203 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.488714 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489169 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489297 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489324 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489430 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489743 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489788 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489798 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.489809 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.490053 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.490086 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.490759 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.490790 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.490276 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.490985 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491020 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-mcd-auth-proxy-config\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491053 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-netns\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491073 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-env-overrides\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491091 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cc616aab-babf-4203-823a-6d5e9d5edcd3-hosts-file\") pod \"node-resolver-vf7m4\" (UID: \"cc616aab-babf-4203-823a-6d5e9d5edcd3\") " pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491110 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-socket-dir-parent\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491129 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-kubelet\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491150 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-etc-kubernetes\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491169 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7abc7c60-6863-4f05-b6d3-09bacb20a405-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491192 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9dxz\" (UniqueName: \"kubernetes.io/projected/5ec6607b-f406-49cd-84db-c83837e9735b-kube-api-access-t9dxz\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491209 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-slash\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491229 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-var-lib-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491256 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491278 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/feb53af0-677a-4160-a671-f183573c68da-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491298 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-k8s-cni-cncf-io\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491316 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-multus-certs\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.491337 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.492134 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.492789 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.492820 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.492908 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493072 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493071 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.493190 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493186 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-bin\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.493192 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493222 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.493268 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:21.993247383 +0000 UTC m=+123.335719781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493297 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-cnibin\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.493333 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:21.993309225 +0000 UTC m=+123.335781823 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493377 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493548 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493594 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493724 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493857 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2nbj\" (UniqueName: \"kubernetes.io/projected/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-kube-api-access-c2nbj\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493891 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.493955 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.494243 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.494553 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.494839 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.494394 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/80b0e746-0802-4112-b77b-4db85dc85a9c-ovn-node-metrics-cert\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495038 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-system-cni-dir\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495085 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-cni-bin\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495180 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495319 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495340 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495357 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495372 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495389 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495399 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495410 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495422 4750 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495433 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495449 4750 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495465 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495544 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495563 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495578 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495590 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495603 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495524 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495689 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.495714 4750 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496049 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496220 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496445 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496514 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496529 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496543 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496555 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496569 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.496692 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497060 4750 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497081 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497122 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497145 4750 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497203 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497227 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497249 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497267 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497285 4750 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497295 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497320 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497354 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497373 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497386 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497400 4750 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497416 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497432 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497445 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497459 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497470 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497483 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497495 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497506 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497522 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497535 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497579 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497610 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.497679 4750 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498203 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.498239 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:27:21.998192994 +0000 UTC m=+123.340665392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498265 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498284 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498295 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498309 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498319 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498314 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498328 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498341 4750 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498350 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498379 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498400 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498428 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498441 4750 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498452 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498463 4750 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498475 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498488 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498499 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498510 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498522 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498532 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498542 4750 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498552 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498565 4750 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498574 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498583 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498593 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498601 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498612 4750 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498647 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498658 4750 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498667 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498678 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498688 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498697 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498707 4750 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498717 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498742 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498754 4750 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498764 4750 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498774 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498788 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498786 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498798 4750 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498864 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498886 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498903 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498918 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499022 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499134 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499413 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.498938 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499456 4750 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499466 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499476 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499487 4750 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499498 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499508 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499519 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499529 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499546 4750 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499557 4750 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499567 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499577 4750 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499591 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499611 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499624 4750 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499651 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499664 4750 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499676 4750 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499690 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499704 4750 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499716 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499728 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499740 4750 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499753 4750 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499765 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499777 4750 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499789 4750 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499800 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499813 4750 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499824 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499835 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499848 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499859 4750 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499871 4750 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499886 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499901 4750 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499914 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499924 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499936 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499948 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499964 4750 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499976 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.499987 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500002 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500014 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500028 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500041 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500052 4750 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500064 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500076 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500088 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500099 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500110 4750 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500120 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500131 4750 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500143 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.500154 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.503902 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.505897 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.506105 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.507016 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.507057 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.507077 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.507194 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:22.007133703 +0000 UTC m=+123.349606301 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.513707 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.514318 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.514822 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.514889 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.514900 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.514941 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.518229 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.518239 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.518380 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.518407 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.518431 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.518424 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.518493 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:22.018468844 +0000 UTC m=+123.360941432 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.518501 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.520842 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.521121 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.522173 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.522932 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.522985 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.523324 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.523331 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.523447 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.523494 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.524348 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.524760 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.524945 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.525008 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.525344 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.525355 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.525473 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.530517 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.539070 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.547614 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.550186 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.561282 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.573282 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.586065 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.599038 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601671 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-cnibin\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601729 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7abc7c60-6863-4f05-b6d3-09bacb20a405-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601753 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ec6607b-f406-49cd-84db-c83837e9735b-serviceca\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601772 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-hostroot\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601799 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-daemon-config\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601818 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-cnibin\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601831 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-etc-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601889 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sw7v\" (UniqueName: \"kubernetes.io/projected/cc616aab-babf-4203-823a-6d5e9d5edcd3-kube-api-access-2sw7v\") pod \"node-resolver-vf7m4\" (UID: \"cc616aab-babf-4203-823a-6d5e9d5edcd3\") " pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601920 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kn78\" (UniqueName: \"kubernetes.io/projected/feb53af0-677a-4160-a671-f183573c68da-kube-api-access-2kn78\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601953 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95qv2\" (UniqueName: \"kubernetes.io/projected/6a32a985-7c9a-4737-86db-d46edf0d0565-kube-api-access-95qv2\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601975 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-kubelet\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601999 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-netd\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.602055 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.602087 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-cni-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.602106 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-rootfs\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.602122 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-system-cni-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.602153 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.602503 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-hostroot\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.602669 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-netd\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.601866 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-etc-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603131 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ec6607b-f406-49cd-84db-c83837e9735b-serviceca\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603271 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-cni-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603274 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-system-cni-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603319 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603354 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-rootfs\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603412 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-kubelet\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603505 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-daemon-config\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603548 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603700 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-node-log\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603725 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-log-socket\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603798 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603819 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-conf-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603761 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-node-log\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603865 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-log-socket\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603953 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-etc-kubernetes\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.603986 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604006 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-mcd-auth-proxy-config\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604023 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-netns\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604040 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-env-overrides\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604056 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cc616aab-babf-4203-823a-6d5e9d5edcd3-hosts-file\") pod \"node-resolver-vf7m4\" (UID: \"cc616aab-babf-4203-823a-6d5e9d5edcd3\") " pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604071 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-socket-dir-parent\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-kubelet\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604103 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7abc7c60-6863-4f05-b6d3-09bacb20a405-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604127 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9dxz\" (UniqueName: \"kubernetes.io/projected/5ec6607b-f406-49cd-84db-c83837e9735b-kube-api-access-t9dxz\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604143 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-slash\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604177 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-var-lib-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604209 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/feb53af0-677a-4160-a671-f183573c68da-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604234 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-k8s-cni-cncf-io\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604253 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-multus-certs\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604270 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-bin\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604312 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-cnibin\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604326 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2nbj\" (UniqueName: \"kubernetes.io/projected/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-kube-api-access-c2nbj\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604357 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604372 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/80b0e746-0802-4112-b77b-4db85dc85a9c-ovn-node-metrics-cert\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604466 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604497 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-netns\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604521 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-conf-dir\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.604551 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-etc-kubernetes\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.604606 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: E0309 18:27:21.604665 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:22.104651415 +0000 UTC m=+123.447123813 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605329 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-mcd-auth-proxy-config\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605371 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605373 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-multus-socket-dir-parent\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605394 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-slash\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605417 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cc616aab-babf-4203-823a-6d5e9d5edcd3-hosts-file\") pod \"node-resolver-vf7m4\" (UID: \"cc616aab-babf-4203-823a-6d5e9d5edcd3\") " pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605440 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-k8s-cni-cncf-io\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605461 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-multus-certs\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605414 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-var-lib-openvswitch\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605493 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-kubelet\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605781 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-env-overrides\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605831 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-bin\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605861 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-cnibin\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.605882 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.606163 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/feb53af0-677a-4160-a671-f183573c68da-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.606284 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-system-cni-dir\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.606843 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7abc7c60-6863-4f05-b6d3-09bacb20a405-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607132 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7abc7c60-6863-4f05-b6d3-09bacb20a405-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607173 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-system-cni-dir\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607241 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-cni-bin\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607257 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fa652827-92a6-4544-b1b6-cba227f8e9f4-cni-binary-copy\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607276 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7abc7c60-6863-4f05-b6d3-09bacb20a405-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607313 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607330 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v5n4\" (UniqueName: \"kubernetes.io/projected/80b0e746-0802-4112-b77b-4db85dc85a9c-kube-api-access-7v5n4\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607359 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-ovn\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607376 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssdqs\" (UniqueName: \"kubernetes.io/projected/fa652827-92a6-4544-b1b6-cba227f8e9f4-kube-api-access-ssdqs\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607430 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-proxy-tls\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607447 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-script-lib\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607462 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-os-release\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607488 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-cni-multus\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607503 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-systemd-units\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607535 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-netns\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607551 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvbbj\" (UniqueName: \"kubernetes.io/projected/7abc7c60-6863-4f05-b6d3-09bacb20a405-kube-api-access-lvbbj\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607566 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607612 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/feb53af0-677a-4160-a671-f183573c68da-cni-binary-copy\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607648 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ec6607b-f406-49cd-84db-c83837e9735b-host\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607665 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-config\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607687 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-os-release\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607744 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607756 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607767 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607781 4750 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607791 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607802 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607812 4750 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607823 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607833 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607850 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607864 4750 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607873 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607882 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607893 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607903 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607912 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607921 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607931 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607941 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607950 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607959 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607971 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607981 4750 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.607990 4750 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608000 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608008 4750 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608020 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608034 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608042 4750 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608055 4750 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608063 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608073 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608082 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608091 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608100 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608111 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608120 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608131 4750 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608175 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/feb53af0-677a-4160-a671-f183573c68da-os-release\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608202 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-script-lib\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608211 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-os-release\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608253 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-cni-bin\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608412 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-var-lib-cni-multus\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608466 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-systemd-units\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608653 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa652827-92a6-4544-b1b6-cba227f8e9f4-host-run-netns\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608717 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fa652827-92a6-4544-b1b6-cba227f8e9f4-cni-binary-copy\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608768 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.608999 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ec6607b-f406-49cd-84db-c83837e9735b-host\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.609140 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7abc7c60-6863-4f05-b6d3-09bacb20a405-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.609161 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/feb53af0-677a-4160-a671-f183573c68da-cni-binary-copy\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.609187 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-ovn\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.612705 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.614349 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.616377 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-config\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.617254 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/80b0e746-0802-4112-b77b-4db85dc85a9c-ovn-node-metrics-cert\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.620336 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-proxy-tls\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.626610 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kn78\" (UniqueName: \"kubernetes.io/projected/feb53af0-677a-4160-a671-f183573c68da-kube-api-access-2kn78\") pod \"multus-additional-cni-plugins-ckgw2\" (UID: \"feb53af0-677a-4160-a671-f183573c68da\") " pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.628217 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvbbj\" (UniqueName: \"kubernetes.io/projected/7abc7c60-6863-4f05-b6d3-09bacb20a405-kube-api-access-lvbbj\") pod \"ovnkube-control-plane-749d76644c-wbb5r\" (UID: \"7abc7c60-6863-4f05-b6d3-09bacb20a405\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.632104 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sw7v\" (UniqueName: \"kubernetes.io/projected/cc616aab-babf-4203-823a-6d5e9d5edcd3-kube-api-access-2sw7v\") pod \"node-resolver-vf7m4\" (UID: \"cc616aab-babf-4203-823a-6d5e9d5edcd3\") " pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.633800 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.635179 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9dxz\" (UniqueName: \"kubernetes.io/projected/5ec6607b-f406-49cd-84db-c83837e9735b-kube-api-access-t9dxz\") pod \"node-ca-4c5gf\" (UID: \"5ec6607b-f406-49cd-84db-c83837e9735b\") " pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.635279 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2nbj\" (UniqueName: \"kubernetes.io/projected/c7f6ff1a-bc64-466a-9f65-59acfede7fc1-kube-api-access-c2nbj\") pod \"machine-config-daemon-pqlpj\" (UID: \"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\") " pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.636818 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95qv2\" (UniqueName: \"kubernetes.io/projected/6a32a985-7c9a-4737-86db-d46edf0d0565-kube-api-access-95qv2\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.637990 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v5n4\" (UniqueName: \"kubernetes.io/projected/80b0e746-0802-4112-b77b-4db85dc85a9c-kube-api-access-7v5n4\") pod \"ovnkube-node-lfwr6\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.641317 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssdqs\" (UniqueName: \"kubernetes.io/projected/fa652827-92a6-4544-b1b6-cba227f8e9f4-kube-api-access-ssdqs\") pod \"multus-vj6kf\" (UID: \"fa652827-92a6-4544-b1b6-cba227f8e9f4\") " pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.647556 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.658841 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.671852 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.683376 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.699669 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.708847 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.717532 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.723678 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.734596 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.742374 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.747519 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-666e9a2a3fcb8cbef1cc34bc22ccd630a0177853c521a1bc4bfe70a050190cde WatchSource:0}: Error finding container 666e9a2a3fcb8cbef1cc34bc22ccd630a0177853c521a1bc4bfe70a050190cde: Status 404 returned error can't find the container with id 666e9a2a3fcb8cbef1cc34bc22ccd630a0177853c521a1bc4bfe70a050190cde Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.751199 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.761372 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.762380 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7f6ff1a_bc64_466a_9f65_59acfede7fc1.slice/crio-92ed2397db4a26e38bec24dd344d390eb9b45875aac6f0952e81a2f67f4882d5 WatchSource:0}: Error finding container 92ed2397db4a26e38bec24dd344d390eb9b45875aac6f0952e81a2f67f4882d5: Status 404 returned error can't find the container with id 92ed2397db4a26e38bec24dd344d390eb9b45875aac6f0952e81a2f67f4882d5 Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.772527 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4c5gf" Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.773954 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-336c23ff284c58641504b9c311fbb706982b897c9c286d962b23c7005316d741 WatchSource:0}: Error finding container 336c23ff284c58641504b9c311fbb706982b897c9c286d962b23c7005316d741: Status 404 returned error can't find the container with id 336c23ff284c58641504b9c311fbb706982b897c9c286d962b23c7005316d741 Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.803227 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80b0e746_0802_4112_b77b_4db85dc85a9c.slice/crio-eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93 WatchSource:0}: Error finding container eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93: Status 404 returned error can't find the container with id eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93 Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.828913 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.840959 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vj6kf" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.845777 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.850423 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vf7m4" Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.851702 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7abc7c60_6863_4f05_b6d3_09bacb20a405.slice/crio-aec66eef9ba444f24edcde860a80bd5659507e2a6623e71f1818ec724e10fa2e WatchSource:0}: Error finding container aec66eef9ba444f24edcde860a80bd5659507e2a6623e71f1818ec724e10fa2e: Status 404 returned error can't find the container with id aec66eef9ba444f24edcde860a80bd5659507e2a6623e71f1818ec724e10fa2e Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.871545 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeb53af0_677a_4160_a671_f183573c68da.slice/crio-2b7742a8c844638bed0eafc29752a33ffe1f218712b15a027650da9a25a9aabc WatchSource:0}: Error finding container 2b7742a8c844638bed0eafc29752a33ffe1f218712b15a027650da9a25a9aabc: Status 404 returned error can't find the container with id 2b7742a8c844638bed0eafc29752a33ffe1f218712b15a027650da9a25a9aabc Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.879498 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa652827_92a6_4544_b1b6_cba227f8e9f4.slice/crio-bfbd4d584ac8bebecd05df1982859d4b494fef52ecacf0e29d7d230419750cec WatchSource:0}: Error finding container bfbd4d584ac8bebecd05df1982859d4b494fef52ecacf0e29d7d230419750cec: Status 404 returned error can't find the container with id bfbd4d584ac8bebecd05df1982859d4b494fef52ecacf0e29d7d230419750cec Mar 09 18:27:21 crc kubenswrapper[4750]: W0309 18:27:21.889035 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc616aab_babf_4203_823a_6d5e9d5edcd3.slice/crio-b5fcca07de219d6bf082342a248a1397e6abcee2f15047a8fd75c4ee956f4d69 WatchSource:0}: Error finding container b5fcca07de219d6bf082342a248a1397e6abcee2f15047a8fd75c4ee956f4d69: Status 404 returned error can't find the container with id b5fcca07de219d6bf082342a248a1397e6abcee2f15047a8fd75c4ee956f4d69 Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.924551 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"336c23ff284c58641504b9c311fbb706982b897c9c286d962b23c7005316d741"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.929854 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"92ed2397db4a26e38bec24dd344d390eb9b45875aac6f0952e81a2f67f4882d5"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.932185 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"20a857cf0364d5801de3a6d3e4df161ea024ecec1bad69279a08724c6dc0a912"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.933364 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" event={"ID":"7abc7c60-6863-4f05-b6d3-09bacb20a405","Type":"ContainerStarted","Data":"aec66eef9ba444f24edcde860a80bd5659507e2a6623e71f1818ec724e10fa2e"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.934663 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"666e9a2a3fcb8cbef1cc34bc22ccd630a0177853c521a1bc4bfe70a050190cde"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.936611 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4c5gf" event={"ID":"5ec6607b-f406-49cd-84db-c83837e9735b","Type":"ContainerStarted","Data":"3882613c063b5b3493018e8f8553d66b85e1e20e2d67f72b2ca599d9948fc828"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.937420 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerStarted","Data":"bfbd4d584ac8bebecd05df1982859d4b494fef52ecacf0e29d7d230419750cec"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.938215 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerStarted","Data":"2b7742a8c844638bed0eafc29752a33ffe1f218712b15a027650da9a25a9aabc"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.938936 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vf7m4" event={"ID":"cc616aab-babf-4203-823a-6d5e9d5edcd3","Type":"ContainerStarted","Data":"b5fcca07de219d6bf082342a248a1397e6abcee2f15047a8fd75c4ee956f4d69"} Mar 09 18:27:21 crc kubenswrapper[4750]: I0309 18:27:21.940592 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.012973 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013180 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:27:23.013140909 +0000 UTC m=+124.355613447 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.013233 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.013275 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.013309 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013413 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013442 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013502 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:23.013474868 +0000 UTC m=+124.355947266 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013530 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:23.013522899 +0000 UTC m=+124.355995297 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013546 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013586 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013602 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.013703 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:23.013682223 +0000 UTC m=+124.356154611 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.115122 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.115356 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.115299 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.115530 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.115566 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.115598 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:23.115580263 +0000 UTC m=+124.458052661 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.115607 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:22 crc kubenswrapper[4750]: E0309 18:27:22.116818 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:23.116770485 +0000 UTC m=+124.459243053 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.948702 4750 generic.go:334] "Generic (PLEG): container finished" podID="feb53af0-677a-4160-a671-f183573c68da" containerID="a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae" exitCode=0 Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.948789 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerDied","Data":"a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.955035 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" event={"ID":"7abc7c60-6863-4f05-b6d3-09bacb20a405","Type":"ContainerStarted","Data":"66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.955069 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" event={"ID":"7abc7c60-6863-4f05-b6d3-09bacb20a405","Type":"ContainerStarted","Data":"53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.956727 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vf7m4" event={"ID":"cc616aab-babf-4203-823a-6d5e9d5edcd3","Type":"ContainerStarted","Data":"7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.958765 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.958829 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.961305 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" exitCode=0 Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.961525 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.966190 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.966252 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.967709 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.969882 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:22Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.970486 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4c5gf" event={"ID":"5ec6607b-f406-49cd-84db-c83837e9735b","Type":"ContainerStarted","Data":"158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.973935 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerStarted","Data":"a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd"} Mar 09 18:27:22 crc kubenswrapper[4750]: I0309 18:27:22.986677 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:22Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.011850 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.024397 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.024600 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:27:25.024568297 +0000 UTC m=+126.367040695 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.024722 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.024763 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.024829 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.024938 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.024978 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:25.024969938 +0000 UTC m=+126.367442336 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.025398 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.025433 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.025452 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.025522 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:25.025504312 +0000 UTC m=+126.367976710 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.025683 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.025877 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:25.025845211 +0000 UTC m=+126.368317789 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.026315 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.041105 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.068445 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.082596 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.101507 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.114854 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.126341 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.126725 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.128647 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.128923 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.129315 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.129329 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.129291 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:25.129261591 +0000 UTC m=+126.471734179 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.129379 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:25.129363114 +0000 UTC m=+126.471835692 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.137381 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.150815 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.174203 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.186232 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.202212 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.220429 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.234352 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.253163 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.265505 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.278427 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.300053 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.313149 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.325739 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.340986 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.355860 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.370280 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.372521 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.372659 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.372761 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.372835 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.373007 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.373152 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.372999 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:23 crc kubenswrapper[4750]: E0309 18:27:23.373443 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.381936 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.382706 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.383379 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.384068 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.385517 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.386089 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.386716 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.387754 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.388431 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.389445 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.390102 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.391207 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.395093 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.395994 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.397335 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.398483 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.399076 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.400559 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.401163 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.401943 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.403666 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.404538 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.405779 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.406233 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.407522 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.408070 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.408886 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.410467 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.411137 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.411702 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.414118 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.414856 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.415948 4750 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.416064 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.418084 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.419374 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.419959 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.422008 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.422944 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.426767 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.427138 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.427869 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.429029 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.429552 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.430787 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.431494 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.432670 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.433190 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.434204 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.434897 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.436277 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.436920 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.437938 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.438666 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.439237 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.439927 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.440865 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.980438 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.981336 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.981352 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.981364 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} Mar 09 18:27:23 crc kubenswrapper[4750]: I0309 18:27:23.983930 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerStarted","Data":"2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7"} Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.004428 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:23Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.028006 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.042917 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.069963 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.086707 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.101096 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.142137 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.162788 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.177437 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.192208 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.206973 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.225303 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.237672 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.250017 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:24Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:24 crc kubenswrapper[4750]: E0309 18:27:24.477602 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.988158 4750 generic.go:334] "Generic (PLEG): container finished" podID="feb53af0-677a-4160-a671-f183573c68da" containerID="2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7" exitCode=0 Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.988225 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerDied","Data":"2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7"} Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.993863 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.994226 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} Mar 09 18:27:24 crc kubenswrapper[4750]: I0309 18:27:24.995852 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04"} Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.010492 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.028893 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.041847 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.052023 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.052223 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:27:29.0521933 +0000 UTC m=+130.394665708 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.052299 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.052365 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.052393 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.052805 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.052864 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.052884 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.052941 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:29.052927349 +0000 UTC m=+130.395399757 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.053004 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.053034 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:29.053024622 +0000 UTC m=+130.395497040 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.053099 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.053128 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:29.053120404 +0000 UTC m=+130.395592812 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.059865 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.071858 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.087982 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.100893 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.116054 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.128280 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.146921 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.153297 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.153374 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.153549 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.153587 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.153601 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.153677 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:29.153659118 +0000 UTC m=+130.496131516 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.153684 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.153761 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:29.15374387 +0000 UTC m=+130.496216268 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.160830 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.176329 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.190568 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.202355 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.220490 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.235351 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.250793 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.263390 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.278044 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.291071 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.305282 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.316776 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.336394 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.354525 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.373023 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.373185 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.373660 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.373752 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.373839 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.373926 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.373901 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.374062 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:25 crc kubenswrapper[4750]: E0309 18:27:25.374269 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.388570 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.404128 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:25 crc kubenswrapper[4750]: I0309 18:27:25.425593 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:25Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.001646 4750 generic.go:334] "Generic (PLEG): container finished" podID="feb53af0-677a-4160-a671-f183573c68da" containerID="ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f" exitCode=0 Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.001784 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerDied","Data":"ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f"} Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.020047 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.052991 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.069249 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.083940 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.106968 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.120246 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.132365 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.148521 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.166461 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.183135 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.195875 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.209011 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.221266 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.232896 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:26Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.387793 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 18:27:26 crc kubenswrapper[4750]: I0309 18:27:26.388534 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:27:26 crc kubenswrapper[4750]: E0309 18:27:26.388805 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.009017 4750 generic.go:334] "Generic (PLEG): container finished" podID="feb53af0-677a-4160-a671-f183573c68da" containerID="6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85" exitCode=0 Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.009070 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerDied","Data":"6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85"} Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.014455 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.014986 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.015198 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.027942 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.049525 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.064861 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.083545 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.105605 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.122605 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.137580 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.153563 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.165700 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.178880 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.191966 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.203109 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.213372 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.234569 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.245318 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.292490 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.292526 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.292535 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.292550 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.292560 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:27Z","lastTransitionTime":"2026-03-09T18:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.303889 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.307843 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.308017 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.308161 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.308276 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.308367 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:27Z","lastTransitionTime":"2026-03-09T18:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.321008 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.324136 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.324168 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.324179 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.324196 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.324208 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:27Z","lastTransitionTime":"2026-03-09T18:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.335321 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.338595 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.338648 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.338666 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.338680 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.338690 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:27Z","lastTransitionTime":"2026-03-09T18:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.350399 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.353852 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.353912 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.353924 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.353949 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.353970 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:27Z","lastTransitionTime":"2026-03-09T18:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.366862 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:27Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.367038 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.373777 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.373801 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.373923 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.373973 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.374072 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.374149 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:27 crc kubenswrapper[4750]: I0309 18:27:27.374169 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:27 crc kubenswrapper[4750]: E0309 18:27:27.374469 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.022941 4750 generic.go:334] "Generic (PLEG): container finished" podID="feb53af0-677a-4160-a671-f183573c68da" containerID="86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd" exitCode=0 Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.023008 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerDied","Data":"86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd"} Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.035386 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.057197 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.069035 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.126535 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.143310 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.160079 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.173608 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.190100 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.203258 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.219273 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.231503 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.245147 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.260410 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.279083 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:28 crc kubenswrapper[4750]: I0309 18:27:28.295294 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.029841 4750 generic.go:334] "Generic (PLEG): container finished" podID="feb53af0-677a-4160-a671-f183573c68da" containerID="e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9" exitCode=0 Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.029933 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerDied","Data":"e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9"} Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.036236 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7"} Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.036598 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.036678 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.036811 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.051977 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.069542 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.069693 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.070086 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.084770 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.091184 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.091293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.091324 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.091401 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092407 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:27:37.092372006 +0000 UTC m=+138.434844404 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092416 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092494 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092515 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092563 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:37.09255663 +0000 UTC m=+138.435029028 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092567 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092690 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092773 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:37.092751715 +0000 UTC m=+138.435224203 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.092816 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:37.092788326 +0000 UTC m=+138.435260844 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.100797 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.116332 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.132370 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.145223 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.162296 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.177446 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.192248 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.192305 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.192445 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.192456 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.192485 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.192498 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.192519 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:37.192502019 +0000 UTC m=+138.534974417 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.192539 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:37.192528 +0000 UTC m=+138.535000398 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.195949 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.208557 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.227599 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.238192 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.249355 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.264021 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.279183 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.291115 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.306180 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.319285 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.331081 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.341856 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.361120 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.372977 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.373106 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.373150 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.373201 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.372989 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.373287 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.373477 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.373666 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.375741 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.387924 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.403117 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.416783 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.429350 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.442422 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.471982 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: E0309 18:27:29.482042 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.525224 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.545882 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.560063 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.570552 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.584902 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.599924 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.612856 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.626759 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.640862 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.652223 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.663556 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.674091 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.693551 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.706280 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.717490 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:29 crc kubenswrapper[4750]: I0309 18:27:29.730293 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.043694 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" event={"ID":"feb53af0-677a-4160-a671-f183573c68da","Type":"ContainerStarted","Data":"474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f"} Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.057695 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.072719 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.088249 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.098665 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.108144 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.120655 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.132491 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.144753 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.156018 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.172758 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.187225 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.201460 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.216129 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.229397 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.241289 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:30Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:30 crc kubenswrapper[4750]: I0309 18:27:30.680363 4750 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 09 18:27:31 crc kubenswrapper[4750]: I0309 18:27:31.373276 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:31 crc kubenswrapper[4750]: I0309 18:27:31.373302 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:31 crc kubenswrapper[4750]: I0309 18:27:31.373431 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:31 crc kubenswrapper[4750]: E0309 18:27:31.373430 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:31 crc kubenswrapper[4750]: I0309 18:27:31.373474 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:31 crc kubenswrapper[4750]: E0309 18:27:31.373587 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:31 crc kubenswrapper[4750]: E0309 18:27:31.373687 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:31 crc kubenswrapper[4750]: E0309 18:27:31.373747 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.052733 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/0.log" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.058135 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7" exitCode=1 Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.058234 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7"} Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.059704 4750 scope.go:117] "RemoveContainer" containerID="6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.077524 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.095983 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.108254 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.129248 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:31Z\\\",\\\"message\\\":\\\"Slice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0309 18:27:31.865578 6721 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0309 18:27:31.865601 6721 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0309 18:27:31.865666 6721 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0309 18:27:31.865674 6721 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:27:31.865690 6721 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:27:31.865720 6721 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:27:31.865724 6721 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0309 18:27:31.865737 6721 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:27:31.865742 6721 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0309 18:27:31.865756 6721 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:27:31.865775 6721 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:27:31.865780 6721 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:27:31.865799 6721 factory.go:656] Stopping watch factory\\\\nI0309 18:27:31.865807 6721 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:27:31.865817 6721 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:27:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.141983 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.155311 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.170358 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.190969 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.206338 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.219068 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.237071 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.253113 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.268735 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.286668 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:32 crc kubenswrapper[4750]: I0309 18:27:32.301343 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:32Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.063946 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/0.log" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.066347 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8"} Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.066960 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.087314 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.102531 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.118238 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.132821 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.147982 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.175808 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:31Z\\\",\\\"message\\\":\\\"Slice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0309 18:27:31.865578 6721 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0309 18:27:31.865601 6721 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0309 18:27:31.865666 6721 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0309 18:27:31.865674 6721 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:27:31.865690 6721 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:27:31.865720 6721 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:27:31.865724 6721 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0309 18:27:31.865737 6721 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:27:31.865742 6721 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0309 18:27:31.865756 6721 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:27:31.865775 6721 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:27:31.865780 6721 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:27:31.865799 6721 factory.go:656] Stopping watch factory\\\\nI0309 18:27:31.865807 6721 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:27:31.865817 6721 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:27:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.190586 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.202741 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.217753 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.233597 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.246396 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.259851 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.273598 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.289348 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.304016 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.373345 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:33 crc kubenswrapper[4750]: E0309 18:27:33.373520 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.374098 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:33 crc kubenswrapper[4750]: E0309 18:27:33.374171 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.374230 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:33 crc kubenswrapper[4750]: E0309 18:27:33.374290 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:33 crc kubenswrapper[4750]: I0309 18:27:33.374449 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:33 crc kubenswrapper[4750]: E0309 18:27:33.374542 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.072773 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/1.log" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.074335 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/0.log" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.076967 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8" exitCode=1 Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.077040 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8"} Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.077080 4750 scope.go:117] "RemoveContainer" containerID="6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.077692 4750 scope.go:117] "RemoveContainer" containerID="4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8" Mar 09 18:27:34 crc kubenswrapper[4750]: E0309 18:27:34.077861 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.095095 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.109298 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.123015 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.137039 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.158451 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6329f0a14204ebf1b4a6cc8dd0f7f95413caf4d3992cd3914217ae733805f7a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:31Z\\\",\\\"message\\\":\\\"Slice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0309 18:27:31.865578 6721 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0309 18:27:31.865601 6721 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0309 18:27:31.865666 6721 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0309 18:27:31.865674 6721 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:27:31.865690 6721 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:27:31.865720 6721 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:27:31.865724 6721 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0309 18:27:31.865737 6721 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:27:31.865742 6721 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0309 18:27:31.865756 6721 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:27:31.865775 6721 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:27:31.865780 6721 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:27:31.865799 6721 factory.go:656] Stopping watch factory\\\\nI0309 18:27:31.865807 6721 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:27:31.865817 6721 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:27:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:33Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z]\\\\nI0309 18:27:33.073366 6864 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0002b738f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.170133 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.182175 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.201961 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.217274 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.231985 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.249299 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.264620 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.278961 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.290844 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: I0309 18:27:34.306564 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:34Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:34 crc kubenswrapper[4750]: E0309 18:27:34.483814 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.083546 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/1.log" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.089171 4750 scope.go:117] "RemoveContainer" containerID="4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8" Mar 09 18:27:35 crc kubenswrapper[4750]: E0309 18:27:35.089372 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.109975 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.129019 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.143696 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.161535 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.176272 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.192182 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.208234 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.225337 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.240991 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.258602 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.272237 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.286234 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.297995 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.316596 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:33Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z]\\\\nI0309 18:27:33.073366 6864 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0002b738f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.332842 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:35Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.373460 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:35 crc kubenswrapper[4750]: E0309 18:27:35.373672 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.373714 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.373805 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:35 crc kubenswrapper[4750]: E0309 18:27:35.373874 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:35 crc kubenswrapper[4750]: I0309 18:27:35.373714 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:35 crc kubenswrapper[4750]: E0309 18:27:35.374024 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:35 crc kubenswrapper[4750]: E0309 18:27:35.374117 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.190077 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.190205 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.190230 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.190259 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190336 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190385 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:53.19037256 +0000 UTC m=+154.532844958 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190762 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:27:53.19075061 +0000 UTC m=+154.533223008 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190818 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190842 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190859 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190870 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:53.190859622 +0000 UTC m=+154.533332020 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190873 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.190907 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:53.190898334 +0000 UTC m=+154.533370732 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.291085 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.291169 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.291296 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.291314 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.291326 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.291382 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:53.291368486 +0000 UTC m=+154.633840884 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.291454 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.291589 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:27:53.291557471 +0000 UTC m=+154.634029929 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.373404 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.373442 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.373405 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.373428 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.373574 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.373662 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.373786 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.373912 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.398732 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.398762 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.398772 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.398789 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.398800 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:37Z","lastTransitionTime":"2026-03-09T18:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.413966 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:37Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.418352 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.418389 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.418401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.418418 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.418430 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:37Z","lastTransitionTime":"2026-03-09T18:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.435321 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:37Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.439758 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.439798 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.439807 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.439827 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.439838 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:37Z","lastTransitionTime":"2026-03-09T18:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.453886 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:37Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.459285 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.459323 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.459333 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.459351 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.459362 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:37Z","lastTransitionTime":"2026-03-09T18:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.471805 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:37Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.475708 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.475745 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.475758 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.475776 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:37 crc kubenswrapper[4750]: I0309 18:27:37.475789 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:37Z","lastTransitionTime":"2026-03-09T18:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.487815 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:37Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:37 crc kubenswrapper[4750]: E0309 18:27:37.487972 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.372845 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.372852 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.372982 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:39 crc kubenswrapper[4750]: E0309 18:27:39.374149 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.373051 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:39 crc kubenswrapper[4750]: E0309 18:27:39.374316 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:39 crc kubenswrapper[4750]: E0309 18:27:39.374370 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:39 crc kubenswrapper[4750]: E0309 18:27:39.374416 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.390255 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.403835 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.420387 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.432368 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.443555 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.462319 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:33Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z]\\\\nI0309 18:27:33.073366 6864 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0002b738f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.473834 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: E0309 18:27:39.484208 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.488056 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.502210 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.516117 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.529000 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.546763 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.561035 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.574009 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:39 crc kubenswrapper[4750]: I0309 18:27:39.592253 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:40 crc kubenswrapper[4750]: I0309 18:27:40.390668 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 09 18:27:41 crc kubenswrapper[4750]: I0309 18:27:41.372811 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:41 crc kubenswrapper[4750]: I0309 18:27:41.373033 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:41 crc kubenswrapper[4750]: E0309 18:27:41.373237 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:41 crc kubenswrapper[4750]: I0309 18:27:41.373311 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:41 crc kubenswrapper[4750]: I0309 18:27:41.373271 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:41 crc kubenswrapper[4750]: E0309 18:27:41.373551 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:41 crc kubenswrapper[4750]: E0309 18:27:41.373684 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:41 crc kubenswrapper[4750]: E0309 18:27:41.373772 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:41 crc kubenswrapper[4750]: I0309 18:27:41.373947 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:27:41 crc kubenswrapper[4750]: E0309 18:27:41.374099 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:27:43 crc kubenswrapper[4750]: I0309 18:27:43.372929 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:43 crc kubenswrapper[4750]: I0309 18:27:43.372968 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:43 crc kubenswrapper[4750]: I0309 18:27:43.373067 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:43 crc kubenswrapper[4750]: I0309 18:27:43.373169 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:43 crc kubenswrapper[4750]: E0309 18:27:43.373156 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:43 crc kubenswrapper[4750]: E0309 18:27:43.373329 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:43 crc kubenswrapper[4750]: E0309 18:27:43.373474 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:43 crc kubenswrapper[4750]: E0309 18:27:43.373564 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:44 crc kubenswrapper[4750]: E0309 18:27:44.486319 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:45 crc kubenswrapper[4750]: I0309 18:27:45.372420 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:45 crc kubenswrapper[4750]: I0309 18:27:45.372473 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:45 crc kubenswrapper[4750]: I0309 18:27:45.372516 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:45 crc kubenswrapper[4750]: E0309 18:27:45.372659 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:45 crc kubenswrapper[4750]: I0309 18:27:45.372784 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:45 crc kubenswrapper[4750]: E0309 18:27:45.372906 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:45 crc kubenswrapper[4750]: E0309 18:27:45.372980 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:45 crc kubenswrapper[4750]: E0309 18:27:45.373005 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.373358 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.373423 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.373622 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.373702 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.373793 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.373903 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.374040 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.374265 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.539137 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.539194 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.539203 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.539221 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.539231 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:47Z","lastTransitionTime":"2026-03-09T18:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.556333 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:47Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.563617 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.563772 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.563788 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.563809 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.563850 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:47Z","lastTransitionTime":"2026-03-09T18:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.580223 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:47Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.588058 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.588146 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.588188 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.588223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.588250 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:47Z","lastTransitionTime":"2026-03-09T18:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.610848 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:47Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.617165 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.617226 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.617244 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.617270 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.617289 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:47Z","lastTransitionTime":"2026-03-09T18:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.641410 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:47Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.648369 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.648430 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.648454 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.648489 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:47 crc kubenswrapper[4750]: I0309 18:27:47.648512 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:47Z","lastTransitionTime":"2026-03-09T18:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.667396 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:47Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:47 crc kubenswrapper[4750]: E0309 18:27:47.667583 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:27:48 crc kubenswrapper[4750]: I0309 18:27:48.373605 4750 scope.go:117] "RemoveContainer" containerID="4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.151877 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/1.log" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.156421 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821"} Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.156991 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.176349 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.195042 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.221793 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.238417 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.253818 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.270853 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.287554 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.299064 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.313012 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.332784 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:33Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z]\\\\nI0309 18:27:33.073366 6864 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0002b738f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.346782 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.362849 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.372853 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.372965 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.372997 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:49 crc kubenswrapper[4750]: E0309 18:27:49.373190 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.373322 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:49 crc kubenswrapper[4750]: E0309 18:27:49.373464 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:49 crc kubenswrapper[4750]: E0309 18:27:49.373560 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:49 crc kubenswrapper[4750]: E0309 18:27:49.374048 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.379595 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.385915 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.394896 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.413611 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.427259 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.449273 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.469005 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.486067 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: E0309 18:27:49.487237 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.506678 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.523198 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.534953 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.553386 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.565561 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.585988 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.600539 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.614863 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.632322 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.667012 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:33Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z]\\\\nI0309 18:27:33.073366 6864 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0002b738f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.683742 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.703264 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.724836 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:49 crc kubenswrapper[4750]: I0309 18:27:49.740459 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:49Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.163856 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/2.log" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.165557 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/1.log" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.170325 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821" exitCode=1 Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.170415 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821"} Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.170517 4750 scope.go:117] "RemoveContainer" containerID="4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.172216 4750 scope.go:117] "RemoveContainer" containerID="e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821" Mar 09 18:27:50 crc kubenswrapper[4750]: E0309 18:27:50.172554 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.194431 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.216142 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.235768 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.257400 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b5ae7d7749a8287cbff4b1fe44d395c5b3340ccf8c53dc208b8ceb74bfa45d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:33Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:33Z is after 2025-08-24T17:21:41Z]\\\\nI0309 18:27:33.073366 6864 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0002b738f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.274717 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.290132 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.317233 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.337614 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.358086 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.371752 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.388097 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.403072 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.420428 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.436775 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.454954 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.468263 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:50 crc kubenswrapper[4750]: I0309 18:27:50.484474 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:50Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.178025 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/2.log" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.185111 4750 scope.go:117] "RemoveContainer" containerID="e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821" Mar 09 18:27:51 crc kubenswrapper[4750]: E0309 18:27:51.185578 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.202324 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.225658 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.250195 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.266158 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.296821 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.316977 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.337157 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.372265 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.373109 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:51 crc kubenswrapper[4750]: E0309 18:27:51.373270 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.373358 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:51 crc kubenswrapper[4750]: E0309 18:27:51.373419 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.373773 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:51 crc kubenswrapper[4750]: E0309 18:27:51.373858 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.373985 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:51 crc kubenswrapper[4750]: E0309 18:27:51.374073 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.400183 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.415577 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.432399 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.451525 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.472142 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.489683 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.513611 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.531768 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:51 crc kubenswrapper[4750]: I0309 18:27:51.547835 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:51Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.190742 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.190930 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.191085 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:28:25.19105064 +0000 UTC m=+186.533523078 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.292281 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.292551 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:28:25.292514568 +0000 UTC m=+186.634986966 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.292659 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.292726 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.292755 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.292987 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293035 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293055 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293070 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.292991 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293137 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:28:25.293074554 +0000 UTC m=+186.635546992 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293172 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:28:25.293156836 +0000 UTC m=+186.635629274 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.293288 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293306 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293321 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293329 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293383 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:28:25.293345911 +0000 UTC m=+186.635818349 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.293413 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:28:25.293399732 +0000 UTC m=+186.635872160 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.373028 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.373705 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.373119 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.373295 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.374479 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.374289 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:53 crc kubenswrapper[4750]: I0309 18:27:53.373125 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:53 crc kubenswrapper[4750]: E0309 18:27:53.374690 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:54 crc kubenswrapper[4750]: E0309 18:27:54.488311 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:55 crc kubenswrapper[4750]: I0309 18:27:55.372801 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:55 crc kubenswrapper[4750]: I0309 18:27:55.372842 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:55 crc kubenswrapper[4750]: I0309 18:27:55.373004 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:55 crc kubenswrapper[4750]: E0309 18:27:55.373000 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:55 crc kubenswrapper[4750]: I0309 18:27:55.373059 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:55 crc kubenswrapper[4750]: E0309 18:27:55.373215 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:55 crc kubenswrapper[4750]: E0309 18:27:55.373285 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:55 crc kubenswrapper[4750]: I0309 18:27:55.373968 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:27:55 crc kubenswrapper[4750]: E0309 18:27:55.374272 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:27:55 crc kubenswrapper[4750]: E0309 18:27:55.373396 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.373202 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.373316 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.373210 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.373473 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.373227 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.373802 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.373931 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.374105 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.756150 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.756903 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.756926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.756962 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.756982 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:57Z","lastTransitionTime":"2026-03-09T18:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.779426 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:57Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.786375 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.786404 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.786412 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.786431 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.786441 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:57Z","lastTransitionTime":"2026-03-09T18:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.801385 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:57Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.805771 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.805846 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.805866 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.805895 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.805914 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:57Z","lastTransitionTime":"2026-03-09T18:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.826171 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:57Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.832488 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.832532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.832578 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.832602 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.832616 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:57Z","lastTransitionTime":"2026-03-09T18:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.853895 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:57Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.858874 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.858918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.858928 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.858946 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:27:57 crc kubenswrapper[4750]: I0309 18:27:57.858958 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:27:57Z","lastTransitionTime":"2026-03-09T18:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.873868 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:57Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:57 crc kubenswrapper[4750]: E0309 18:27:57.874566 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.372673 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.372803 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:27:59 crc kubenswrapper[4750]: E0309 18:27:59.373725 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.372891 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:27:59 crc kubenswrapper[4750]: E0309 18:27:59.373820 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:27:59 crc kubenswrapper[4750]: E0309 18:27:59.373510 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.372880 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:27:59 crc kubenswrapper[4750]: E0309 18:27:59.373906 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.405054 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.423668 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.440760 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.451810 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.463247 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.479011 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: E0309 18:27:59.489233 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.498481 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.513013 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.528724 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.562355 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.584331 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.605853 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.625386 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.647832 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.664624 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.685380 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:27:59 crc kubenswrapper[4750]: I0309 18:27:59.705300 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:27:59Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:01 crc kubenswrapper[4750]: I0309 18:28:01.372537 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:01 crc kubenswrapper[4750]: I0309 18:28:01.372605 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:01 crc kubenswrapper[4750]: I0309 18:28:01.372756 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:01 crc kubenswrapper[4750]: I0309 18:28:01.372845 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:01 crc kubenswrapper[4750]: E0309 18:28:01.372845 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:01 crc kubenswrapper[4750]: E0309 18:28:01.372992 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:01 crc kubenswrapper[4750]: E0309 18:28:01.373204 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:01 crc kubenswrapper[4750]: E0309 18:28:01.373382 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:02 crc kubenswrapper[4750]: I0309 18:28:02.374186 4750 scope.go:117] "RemoveContainer" containerID="e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821" Mar 09 18:28:02 crc kubenswrapper[4750]: E0309 18:28:02.374473 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:28:02 crc kubenswrapper[4750]: I0309 18:28:02.388583 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 09 18:28:03 crc kubenswrapper[4750]: I0309 18:28:03.372516 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:03 crc kubenswrapper[4750]: I0309 18:28:03.372686 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:03 crc kubenswrapper[4750]: I0309 18:28:03.372474 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:03 crc kubenswrapper[4750]: I0309 18:28:03.372548 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:03 crc kubenswrapper[4750]: E0309 18:28:03.372835 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:03 crc kubenswrapper[4750]: E0309 18:28:03.372909 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:03 crc kubenswrapper[4750]: E0309 18:28:03.373067 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:03 crc kubenswrapper[4750]: E0309 18:28:03.373108 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:04 crc kubenswrapper[4750]: E0309 18:28:04.490653 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:05 crc kubenswrapper[4750]: I0309 18:28:05.372837 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:05 crc kubenswrapper[4750]: I0309 18:28:05.372852 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:05 crc kubenswrapper[4750]: I0309 18:28:05.372886 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:05 crc kubenswrapper[4750]: I0309 18:28:05.372883 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:05 crc kubenswrapper[4750]: E0309 18:28:05.373143 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:05 crc kubenswrapper[4750]: E0309 18:28:05.373245 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:05 crc kubenswrapper[4750]: E0309 18:28:05.373473 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:05 crc kubenswrapper[4750]: E0309 18:28:05.373735 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:06 crc kubenswrapper[4750]: I0309 18:28:06.374589 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:28:06 crc kubenswrapper[4750]: E0309 18:28:06.374983 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:28:07 crc kubenswrapper[4750]: I0309 18:28:07.372699 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:07 crc kubenswrapper[4750]: I0309 18:28:07.372757 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:07 crc kubenswrapper[4750]: I0309 18:28:07.372722 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:07 crc kubenswrapper[4750]: I0309 18:28:07.372713 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:07 crc kubenswrapper[4750]: E0309 18:28:07.372909 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:07 crc kubenswrapper[4750]: E0309 18:28:07.372969 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:07 crc kubenswrapper[4750]: E0309 18:28:07.373035 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:07 crc kubenswrapper[4750]: E0309 18:28:07.373204 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.090604 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.090679 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.090691 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.090708 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.090739 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:08Z","lastTransitionTime":"2026-03-09T18:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:08 crc kubenswrapper[4750]: E0309 18:28:08.115224 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:08Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.121902 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.121948 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.121959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.121978 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.121990 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:08Z","lastTransitionTime":"2026-03-09T18:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:08 crc kubenswrapper[4750]: E0309 18:28:08.176131 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:08Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.184127 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.184181 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.184193 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.184212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.184486 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:08Z","lastTransitionTime":"2026-03-09T18:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:08 crc kubenswrapper[4750]: E0309 18:28:08.210155 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:08Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.214576 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.214650 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.214672 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.214699 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.214716 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:08Z","lastTransitionTime":"2026-03-09T18:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:08 crc kubenswrapper[4750]: E0309 18:28:08.228950 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:08Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.233492 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.233557 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.233566 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.233584 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:08 crc kubenswrapper[4750]: I0309 18:28:08.233595 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:08Z","lastTransitionTime":"2026-03-09T18:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:08 crc kubenswrapper[4750]: E0309 18:28:08.247386 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:08Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:08 crc kubenswrapper[4750]: E0309 18:28:08.247544 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.257570 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/0.log" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.257667 4750 generic.go:334] "Generic (PLEG): container finished" podID="fa652827-92a6-4544-b1b6-cba227f8e9f4" containerID="a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd" exitCode=1 Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.257713 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerDied","Data":"a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd"} Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.258283 4750 scope.go:117] "RemoveContainer" containerID="a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.288109 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.304272 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.318505 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.333163 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.343217 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.359750 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.372622 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.372686 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.372830 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:09 crc kubenswrapper[4750]: E0309 18:28:09.372897 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:09 crc kubenswrapper[4750]: E0309 18:28:09.373105 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:09 crc kubenswrapper[4750]: E0309 18:28:09.373420 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.373705 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.374255 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: E0309 18:28:09.374816 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.394659 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.407599 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.432001 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.443660 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.465109 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.480482 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: E0309 18:28:09.491484 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.491612 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.506762 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.516466 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.527129 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.540803 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.553775 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.566743 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.581465 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.595883 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.608418 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.619539 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.629551 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.653872 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.666918 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.678237 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.692013 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.705667 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.717358 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.732506 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.748949 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.759785 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.774843 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:09 crc kubenswrapper[4750]: I0309 18:28:09.788917 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:09Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.264075 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/0.log" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.264421 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerStarted","Data":"2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c"} Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.281956 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.301484 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.318947 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.336001 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.349533 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.377507 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.388138 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.388511 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.412160 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.427843 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.439915 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.454033 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.466018 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.477595 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.492267 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.510273 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.528047 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.543706 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:10 crc kubenswrapper[4750]: I0309 18:28:10.561223 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:10Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:11 crc kubenswrapper[4750]: I0309 18:28:11.373437 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:11 crc kubenswrapper[4750]: I0309 18:28:11.373587 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:11 crc kubenswrapper[4750]: I0309 18:28:11.373437 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:11 crc kubenswrapper[4750]: I0309 18:28:11.373754 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:11 crc kubenswrapper[4750]: E0309 18:28:11.373692 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:11 crc kubenswrapper[4750]: E0309 18:28:11.373956 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:11 crc kubenswrapper[4750]: E0309 18:28:11.374093 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:11 crc kubenswrapper[4750]: E0309 18:28:11.374242 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:13 crc kubenswrapper[4750]: I0309 18:28:13.372721 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:13 crc kubenswrapper[4750]: I0309 18:28:13.372779 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:13 crc kubenswrapper[4750]: I0309 18:28:13.372883 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:13 crc kubenswrapper[4750]: I0309 18:28:13.372929 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:13 crc kubenswrapper[4750]: E0309 18:28:13.373008 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:13 crc kubenswrapper[4750]: E0309 18:28:13.373176 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:13 crc kubenswrapper[4750]: E0309 18:28:13.373365 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:13 crc kubenswrapper[4750]: E0309 18:28:13.373537 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:14 crc kubenswrapper[4750]: E0309 18:28:14.493152 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:15 crc kubenswrapper[4750]: I0309 18:28:15.372490 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:15 crc kubenswrapper[4750]: I0309 18:28:15.372616 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:15 crc kubenswrapper[4750]: I0309 18:28:15.372524 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:15 crc kubenswrapper[4750]: E0309 18:28:15.372720 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:15 crc kubenswrapper[4750]: I0309 18:28:15.372776 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:15 crc kubenswrapper[4750]: E0309 18:28:15.372796 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:15 crc kubenswrapper[4750]: E0309 18:28:15.372854 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:15 crc kubenswrapper[4750]: E0309 18:28:15.373594 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:15 crc kubenswrapper[4750]: I0309 18:28:15.374311 4750 scope.go:117] "RemoveContainer" containerID="e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.287108 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/2.log" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.289911 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.290695 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.306625 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.319145 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.332498 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.344287 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.361199 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5838f42d-a451-4b9f-b733-3364a6357b5d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1577b46bd60cd09246da1ba3c6d76c6da8782969bbefe301ded194092da54f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e617cc775c5b425a56aacf7fb7a049a8117366cee205d7289d896d04386608a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://899a0632bb66ad4c1454d05ab958515d304030a7fa0ddd190d9a28133e4af8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://531e2e851c9d42b4d3dc21665d1d36326a0c63d6cef342c9560c26590f8256b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171f973093d7d57ceae547565c16f8a12f0889ad5d0ad2b870320faad0e1cef4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.374895 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.387216 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.407771 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.423831 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.436746 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.447361 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.473540 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.487280 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.501525 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.517736 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.532000 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.547378 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.560674 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:16 crc kubenswrapper[4750]: I0309 18:28:16.575481 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:16Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.297254 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/3.log" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.298497 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/2.log" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.303727 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" exitCode=1 Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.303803 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.303875 4750 scope.go:117] "RemoveContainer" containerID="e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.305233 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:28:17 crc kubenswrapper[4750]: E0309 18:28:17.305732 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.350075 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5838f42d-a451-4b9f-b733-3364a6357b5d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1577b46bd60cd09246da1ba3c6d76c6da8782969bbefe301ded194092da54f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e617cc775c5b425a56aacf7fb7a049a8117366cee205d7289d896d04386608a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://899a0632bb66ad4c1454d05ab958515d304030a7fa0ddd190d9a28133e4af8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://531e2e851c9d42b4d3dc21665d1d36326a0c63d6cef342c9560c26590f8256b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171f973093d7d57ceae547565c16f8a12f0889ad5d0ad2b870320faad0e1cef4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.372420 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:17 crc kubenswrapper[4750]: E0309 18:28:17.372724 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.372800 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.372850 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:17 crc kubenswrapper[4750]: E0309 18:28:17.373001 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:17 crc kubenswrapper[4750]: E0309 18:28:17.373133 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.373291 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:17 crc kubenswrapper[4750]: E0309 18:28:17.373425 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.376839 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.405354 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.425723 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.443354 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.463406 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.480467 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.499856 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.520394 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.537667 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.557932 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.575393 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.606700 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e311e24c03b6b6a4278a9fab9ddc2bf60fe871b3694001d98e33f38684e821\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:27:49Z\\\",\\\"message\\\":\\\".go:365] Adding new object: *v1.Pod openshift-multus/multus-vj6kf\\\\nI0309 18:27:49.456920 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0309 18:27:49.456925 7045 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0309 18:27:49.456913 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0309 18:27:49.456860 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456944 7045 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-4c5gf\\\\nI0309 18:27:49.456962 7045 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-4c5gf in node crc\\\\nI0309 18:27:49.456810 7045 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0309 18:27:49.456970 7045 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:16Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:28:16.291742 7363 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0309 18:28:16.291751 7363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0309 18:28:16.291766 7363 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:28:16.291771 7363 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:28:16.291783 7363 handler.go:208] Removed *v1.Node event handler 7\\\\nI0309 18:28:16.291798 7363 handler.go:208] Removed *v1.Node event handler 2\\\\nI0309 18:28:16.291840 7363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:28:16.291846 7363 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:28:16.291869 7363 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:28:16.291878 7363 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:28:16.291880 7363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:28:16.291988 7363 factory.go:656] Stopping watch factory\\\\nI0309 18:28:16.292060 7363 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:28:16.292004 7363 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0309 18:28:16.292134 7363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0309 18:28:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.625136 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.640735 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.663388 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.681083 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.703585 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:17 crc kubenswrapper[4750]: I0309 18:28:17.721598 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:17Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.311151 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/3.log" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.316179 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:28:18 crc kubenswrapper[4750]: E0309 18:28:18.316392 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.342073 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.365102 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.379688 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.393023 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.406423 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.406474 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.406483 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.406499 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.406512 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:18Z","lastTransitionTime":"2026-03-09T18:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:18 crc kubenswrapper[4750]: E0309 18:28:18.423224 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.426601 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:16Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:28:16.291742 7363 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0309 18:28:16.291751 7363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0309 18:28:16.291766 7363 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:28:16.291771 7363 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:28:16.291783 7363 handler.go:208] Removed *v1.Node event handler 7\\\\nI0309 18:28:16.291798 7363 handler.go:208] Removed *v1.Node event handler 2\\\\nI0309 18:28:16.291840 7363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:28:16.291846 7363 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:28:16.291869 7363 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:28:16.291878 7363 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:28:16.291880 7363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:28:16.291988 7363 factory.go:656] Stopping watch factory\\\\nI0309 18:28:16.292060 7363 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:28:16.292004 7363 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0309 18:28:16.292134 7363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0309 18:28:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.428002 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.428037 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.428047 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.428061 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.428073 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:18Z","lastTransitionTime":"2026-03-09T18:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.441234 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: E0309 18:28:18.445466 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.449803 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.449862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.449879 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.449904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.449950 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:18Z","lastTransitionTime":"2026-03-09T18:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.455008 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: E0309 18:28:18.465988 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.470616 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.470700 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.470724 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.470749 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.470767 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:18Z","lastTransitionTime":"2026-03-09T18:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.470935 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: E0309 18:28:18.486855 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.488334 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.491266 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.491322 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.491355 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.491388 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.491411 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:18Z","lastTransitionTime":"2026-03-09T18:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.502632 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: E0309 18:28:18.506143 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: E0309 18:28:18.506340 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.522267 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5838f42d-a451-4b9f-b733-3364a6357b5d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1577b46bd60cd09246da1ba3c6d76c6da8782969bbefe301ded194092da54f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e617cc775c5b425a56aacf7fb7a049a8117366cee205d7289d896d04386608a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://899a0632bb66ad4c1454d05ab958515d304030a7fa0ddd190d9a28133e4af8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://531e2e851c9d42b4d3dc21665d1d36326a0c63d6cef342c9560c26590f8256b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171f973093d7d57ceae547565c16f8a12f0889ad5d0ad2b870320faad0e1cef4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.537024 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.547078 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.569186 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.587776 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.601122 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.617539 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.629401 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:18 crc kubenswrapper[4750]: I0309 18:28:18.645449 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:18Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.372353 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.372491 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:19 crc kubenswrapper[4750]: E0309 18:28:19.372610 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:19 crc kubenswrapper[4750]: E0309 18:28:19.372802 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.374054 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.375070 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:28:19 crc kubenswrapper[4750]: E0309 18:28:19.375389 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 18:28:19 crc kubenswrapper[4750]: E0309 18:28:19.375918 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.376170 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:19 crc kubenswrapper[4750]: E0309 18:28:19.376394 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.390168 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.412013 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.428612 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.445018 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.462266 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: E0309 18:28:19.494239 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.497679 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:16Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:28:16.291742 7363 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0309 18:28:16.291751 7363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0309 18:28:16.291766 7363 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:28:16.291771 7363 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:28:16.291783 7363 handler.go:208] Removed *v1.Node event handler 7\\\\nI0309 18:28:16.291798 7363 handler.go:208] Removed *v1.Node event handler 2\\\\nI0309 18:28:16.291840 7363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:28:16.291846 7363 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:28:16.291869 7363 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:28:16.291878 7363 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:28:16.291880 7363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:28:16.291988 7363 factory.go:656] Stopping watch factory\\\\nI0309 18:28:16.292060 7363 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:28:16.292004 7363 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0309 18:28:16.292134 7363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0309 18:28:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.516738 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.537704 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.556298 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.575040 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.606977 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5838f42d-a451-4b9f-b733-3364a6357b5d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1577b46bd60cd09246da1ba3c6d76c6da8782969bbefe301ded194092da54f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e617cc775c5b425a56aacf7fb7a049a8117366cee205d7289d896d04386608a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://899a0632bb66ad4c1454d05ab958515d304030a7fa0ddd190d9a28133e4af8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://531e2e851c9d42b4d3dc21665d1d36326a0c63d6cef342c9560c26590f8256b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171f973093d7d57ceae547565c16f8a12f0889ad5d0ad2b870320faad0e1cef4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.623962 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.635433 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.647082 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.673371 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.688523 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.701854 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.714119 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:19 crc kubenswrapper[4750]: I0309 18:28:19.727921 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:19Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:21 crc kubenswrapper[4750]: I0309 18:28:21.372602 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:21 crc kubenswrapper[4750]: I0309 18:28:21.372701 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:21 crc kubenswrapper[4750]: I0309 18:28:21.372748 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:21 crc kubenswrapper[4750]: I0309 18:28:21.372726 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:21 crc kubenswrapper[4750]: E0309 18:28:21.372926 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:21 crc kubenswrapper[4750]: E0309 18:28:21.372987 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:21 crc kubenswrapper[4750]: E0309 18:28:21.373152 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:21 crc kubenswrapper[4750]: E0309 18:28:21.373520 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:23 crc kubenswrapper[4750]: I0309 18:28:23.373343 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:23 crc kubenswrapper[4750]: I0309 18:28:23.373343 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:23 crc kubenswrapper[4750]: E0309 18:28:23.373541 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:23 crc kubenswrapper[4750]: I0309 18:28:23.373376 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:23 crc kubenswrapper[4750]: E0309 18:28:23.373741 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:23 crc kubenswrapper[4750]: I0309 18:28:23.373825 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:23 crc kubenswrapper[4750]: E0309 18:28:23.373930 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:23 crc kubenswrapper[4750]: E0309 18:28:23.374037 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:24 crc kubenswrapper[4750]: E0309 18:28:24.495573 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.195532 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.195809 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.196181 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.196150053 +0000 UTC m=+250.538622451 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.297487 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.297746 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.297852 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.297789597 +0000 UTC m=+250.640262025 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.297967 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.297993 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298016 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298084 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.298065064 +0000 UTC m=+250.640537492 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.297921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.298192 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.298243 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298460 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298542 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.298524026 +0000 UTC m=+250.640996464 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298578 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298675 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298695 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298788 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.298764822 +0000 UTC m=+250.641237240 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298795 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.298870 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs podName:6a32a985-7c9a-4737-86db-d46edf0d0565 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.298850024 +0000 UTC m=+250.641322452 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs") pod "network-metrics-daemon-wl7w6" (UID: "6a32a985-7c9a-4737-86db-d46edf0d0565") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.374007 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.374071 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.374065 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:25 crc kubenswrapper[4750]: I0309 18:28:25.374156 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.374417 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.374694 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.374796 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:25 crc kubenswrapper[4750]: E0309 18:28:25.374969 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:27 crc kubenswrapper[4750]: I0309 18:28:27.372729 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:27 crc kubenswrapper[4750]: I0309 18:28:27.372879 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:27 crc kubenswrapper[4750]: I0309 18:28:27.372947 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:27 crc kubenswrapper[4750]: E0309 18:28:27.372988 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:27 crc kubenswrapper[4750]: I0309 18:28:27.373007 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:27 crc kubenswrapper[4750]: E0309 18:28:27.373133 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:27 crc kubenswrapper[4750]: E0309 18:28:27.373360 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:27 crc kubenswrapper[4750]: E0309 18:28:27.373430 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.626586 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.626674 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.626691 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.626713 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.626733 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:28Z","lastTransitionTime":"2026-03-09T18:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:28 crc kubenswrapper[4750]: E0309 18:28:28.648680 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.654282 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.654338 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.654397 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.654438 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.654466 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:28Z","lastTransitionTime":"2026-03-09T18:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:28 crc kubenswrapper[4750]: E0309 18:28:28.672679 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.677785 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.677846 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.677862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.677896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.677918 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:28Z","lastTransitionTime":"2026-03-09T18:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:28 crc kubenswrapper[4750]: E0309 18:28:28.700768 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.705844 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.705911 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.705931 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.705966 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.705987 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:28Z","lastTransitionTime":"2026-03-09T18:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:28 crc kubenswrapper[4750]: E0309 18:28:28.731128 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.735495 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.735546 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.735562 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.735587 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:28 crc kubenswrapper[4750]: I0309 18:28:28.735602 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:28Z","lastTransitionTime":"2026-03-09T18:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:28 crc kubenswrapper[4750]: E0309 18:28:28.753076 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:28Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:28 crc kubenswrapper[4750]: E0309 18:28:28.753236 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.373070 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:29 crc kubenswrapper[4750]: E0309 18:28:29.373405 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.373473 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.373522 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.373964 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:29 crc kubenswrapper[4750]: E0309 18:28:29.374122 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:29 crc kubenswrapper[4750]: E0309 18:28:29.374263 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:29 crc kubenswrapper[4750]: E0309 18:28:29.374319 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.374551 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:28:29 crc kubenswrapper[4750]: E0309 18:28:29.374777 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.395412 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.407003 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.425999 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.449147 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.461772 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.476858 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.490812 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: E0309 18:28:29.496087 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.513247 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5838f42d-a451-4b9f-b733-3364a6357b5d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1577b46bd60cd09246da1ba3c6d76c6da8782969bbefe301ded194092da54f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e617cc775c5b425a56aacf7fb7a049a8117366cee205d7289d896d04386608a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://899a0632bb66ad4c1454d05ab958515d304030a7fa0ddd190d9a28133e4af8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://531e2e851c9d42b4d3dc21665d1d36326a0c63d6cef342c9560c26590f8256b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171f973093d7d57ceae547565c16f8a12f0889ad5d0ad2b870320faad0e1cef4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.529298 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.547085 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.564581 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.581739 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.597337 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.612566 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.641749 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:16Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:28:16.291742 7363 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0309 18:28:16.291751 7363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0309 18:28:16.291766 7363 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:28:16.291771 7363 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:28:16.291783 7363 handler.go:208] Removed *v1.Node event handler 7\\\\nI0309 18:28:16.291798 7363 handler.go:208] Removed *v1.Node event handler 2\\\\nI0309 18:28:16.291840 7363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:28:16.291846 7363 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:28:16.291869 7363 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:28:16.291878 7363 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:28:16.291880 7363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:28:16.291988 7363 factory.go:656] Stopping watch factory\\\\nI0309 18:28:16.292060 7363 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:28:16.292004 7363 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0309 18:28:16.292134 7363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0309 18:28:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.654886 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.667910 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.686119 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:29 crc kubenswrapper[4750]: I0309 18:28:29.703772 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:29Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:31 crc kubenswrapper[4750]: I0309 18:28:31.373325 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:31 crc kubenswrapper[4750]: I0309 18:28:31.373380 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:31 crc kubenswrapper[4750]: I0309 18:28:31.373332 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:31 crc kubenswrapper[4750]: I0309 18:28:31.373553 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:31 crc kubenswrapper[4750]: E0309 18:28:31.373796 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:31 crc kubenswrapper[4750]: E0309 18:28:31.374007 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:31 crc kubenswrapper[4750]: E0309 18:28:31.374176 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:31 crc kubenswrapper[4750]: E0309 18:28:31.374355 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:32 crc kubenswrapper[4750]: I0309 18:28:32.373245 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.372697 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.372735 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.372697 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:33 crc kubenswrapper[4750]: E0309 18:28:33.372876 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.372872 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:33 crc kubenswrapper[4750]: E0309 18:28:33.373053 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:33 crc kubenswrapper[4750]: E0309 18:28:33.373407 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:33 crc kubenswrapper[4750]: E0309 18:28:33.373733 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.524929 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.527220 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14"} Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.527504 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.545381 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9dcff8eb-759d-411a-8fd8-51678f9c96fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://370dafd6687dcc25e06862f4603876cc9876756c500fa03b1e11b15acef0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e454c5108753c59257575dfe723085a74c986efd42e4135ad036670ea175050\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:26:20Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0309 18:25:51.093583 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0309 18:25:51.095252 1 observer_polling.go:159] Starting file observer\\\\nI0309 18:25:51.098538 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0309 18:25:51.099673 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0309 18:26:16.562176 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0309 18:26:20.693916 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0309 18:26:20.694022 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:51Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdf57ea8db28a5f72d4eeb1206e36c78de7920005e52d95daa97a7859bf00ac5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e050562a09e996cb849aab6c77bac7bd5a628820a9b482ba19ab7f891d6b94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.562707 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.574168 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e7f6411702370dfc09d4c7e7c7b29e113d3bcda4969778de2d6908474e08e04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.588263 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feb53af0-677a-4160-a671-f183573c68da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474ad9e75f8e1c4c0607c63b41e418f58ff0148a6719587b3b2f8cceb22f3a3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3bbfc782086e44be53b6ece5557a4f74db7f4d75525b3af7563861033e093ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e41f25356a7488e4b661bed883f12e51de0d830709e8c2b8303be26739b12b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba85c567b8f482e7dd5b18a744dc7b5d56080bd8604c242c090e24b30fdb112f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6877f13cfd57b8a075a7503dca95d85fdff660908bb2e3976b573433fc1d6c85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86ffeb01265bef993d758d849d59e26fc3f92cee993f01ef31e7b6e6f38d0fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b1f9e6ca6a66b01787c18db5404f05d9ab4d4f44f204595494ebb69b190bf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kn78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ckgw2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.601496 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vj6kf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa652827-92a6-4544-b1b6-cba227f8e9f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:08Z\\\",\\\"message\\\":\\\"2026-03-09T18:27:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274\\\\n2026-03-09T18:27:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87c754c4-2fa7-426a-9827-a14750a9d274 to /host/opt/cni/bin/\\\\n2026-03-09T18:27:23Z [verbose] multus-daemon started\\\\n2026-03-09T18:27:23Z [verbose] Readiness Indicator file check\\\\n2026-03-09T18:28:08Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssdqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vj6kf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.613844 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.629163 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7f6ff1a-bc64-466a-9f65-59acfede7fc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47a0e68f574b0c69bc18543dfb28fb9448891aa44c149c175d582ab6679af8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2nbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pqlpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.638611 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.655141 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:16Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:28:16.291742 7363 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0309 18:28:16.291751 7363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0309 18:28:16.291766 7363 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:28:16.291771 7363 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:28:16.291783 7363 handler.go:208] Removed *v1.Node event handler 7\\\\nI0309 18:28:16.291798 7363 handler.go:208] Removed *v1.Node event handler 2\\\\nI0309 18:28:16.291840 7363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:28:16.291846 7363 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:28:16.291869 7363 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:28:16.291878 7363 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:28:16.291880 7363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:28:16.291988 7363 factory.go:656] Stopping watch factory\\\\nI0309 18:28:16.292060 7363 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:28:16.292004 7363 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0309 18:28:16.292134 7363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0309 18:28:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.666477 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.681978 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.698253 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db02cf445c2412cecb712172ac4210d0dd7d00469a86492d05d841f7eaba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.712520 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.724976 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7abc7c60-6863-4f05-b6d3-09bacb20a405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a4f8f2fa9f9e6f407b9cab689e978310bef2fc6edb25f28b5fac117561c81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f2952eedae1c92a3ae861da113487ac2724c7a637e68e37440110c0f32efae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvbbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wbb5r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.742998 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5838f42d-a451-4b9f-b733-3364a6357b5d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1577b46bd60cd09246da1ba3c6d76c6da8782969bbefe301ded194092da54f70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e617cc775c5b425a56aacf7fb7a049a8117366cee205d7289d896d04386608a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://899a0632bb66ad4c1454d05ab958515d304030a7fa0ddd190d9a28133e4af8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://531e2e851c9d42b4d3dc21665d1d36326a0c63d6cef342c9560c26590f8256b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171f973093d7d57ceae547565c16f8a12f0889ad5d0ad2b870320faad0e1cef4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ffa9e2fffd9dbf24a45552cdef86a4e6be071a007a0be4ed37c2ce79814119\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814424b52e62520f5241caa18e971e6d8ef81a57db982b6d7e17b0fc32552eb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4793f9c5927a56bc22fd636f0a517e0febab3ed14f8c38b66ffb9a1c306c53e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.756466 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T18:27:06Z\\\",\\\"message\\\":\\\"file observer\\\\nW0309 18:27:06.080404 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 18:27:06.080570 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 18:27:06.081498 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-754369617/tls.crt::/tmp/serving-cert-754369617/tls.key\\\\\\\"\\\\nI0309 18:27:06.521709 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 18:27:06.524558 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 18:27:06.524580 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 18:27:06.524605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 18:27:06.524611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 18:27:06.532243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 18:27:06.532293 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532300 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 18:27:06.532308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 18:27:06.532312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 18:27:06.532316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 18:27:06.532319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 18:27:06.532259 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 18:27:06.534714 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:28:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.767694 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9335add1-4399-4b3b-b14c-8055408b4349\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e0e7600d555d12d3156444eb18b96a67826e9fb08cff116ab2e57aef2381b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d92b90a06e0ea0884d48f19deee8249e4cde144fa85cfb51c768f33a00e77f26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.778569 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f7c2f7-2dd9-4a05-8025-8a7954b36f13\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:26:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd92a5fad9d422142b3e78ae3eb32b448024c18650deaab65f86ae3614dd87d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b1a87eb072f2211e4c0c515e221e571457c1eecab31e74ae14133dafdc8a12c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://883b6580d985f290759ec06f77a88018266f6d706bb0d54d256b3fa65c6174f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ba10c8b12092467c295061012f41277d0b4b15124a1840aa81c16feec9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:25:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:25:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:25:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:33 crc kubenswrapper[4750]: I0309 18:28:33.792748 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aac08b30ea0949778d25a1ce37dbed6d7b98f971bd8296248f1118a3b218794a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46219f716968e9eb87403b0971cf7fbd4f9e93bf5099e5f3c9400fca5c36a9bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:33Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:34 crc kubenswrapper[4750]: E0309 18:28:34.497726 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:35 crc kubenswrapper[4750]: I0309 18:28:35.372315 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:35 crc kubenswrapper[4750]: I0309 18:28:35.372413 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:35 crc kubenswrapper[4750]: E0309 18:28:35.372511 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:35 crc kubenswrapper[4750]: E0309 18:28:35.372677 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:35 crc kubenswrapper[4750]: I0309 18:28:35.372797 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:35 crc kubenswrapper[4750]: I0309 18:28:35.372927 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:35 crc kubenswrapper[4750]: E0309 18:28:35.373098 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:35 crc kubenswrapper[4750]: E0309 18:28:35.373148 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:37 crc kubenswrapper[4750]: I0309 18:28:37.373392 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:37 crc kubenswrapper[4750]: I0309 18:28:37.373436 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:37 crc kubenswrapper[4750]: E0309 18:28:37.374054 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:37 crc kubenswrapper[4750]: I0309 18:28:37.373535 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:37 crc kubenswrapper[4750]: I0309 18:28:37.373498 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:37 crc kubenswrapper[4750]: E0309 18:28:37.374220 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:37 crc kubenswrapper[4750]: E0309 18:28:37.374350 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:37 crc kubenswrapper[4750]: E0309 18:28:37.374446 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.865823 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.865879 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.865891 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.865911 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.865923 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:38Z","lastTransitionTime":"2026-03-09T18:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:38 crc kubenswrapper[4750]: E0309 18:28:38.885443 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:38Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.890933 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.891188 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.891348 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.891518 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.891710 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:38Z","lastTransitionTime":"2026-03-09T18:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:38 crc kubenswrapper[4750]: E0309 18:28:38.908789 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:38Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.915201 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.915272 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.915297 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.915330 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.915352 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:38Z","lastTransitionTime":"2026-03-09T18:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:38 crc kubenswrapper[4750]: E0309 18:28:38.931559 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:38Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.936107 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.936263 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.936333 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.936404 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.936488 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:38Z","lastTransitionTime":"2026-03-09T18:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:38 crc kubenswrapper[4750]: E0309 18:28:38.952390 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:38Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.957183 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.957223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.957235 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.957259 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:38 crc kubenswrapper[4750]: I0309 18:28:38.957273 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:38Z","lastTransitionTime":"2026-03-09T18:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:38 crc kubenswrapper[4750]: E0309 18:28:38.974560 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T18:28:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"836d638e-2eed-4bd3-92fd-26e5289ded19\\\",\\\"systemUUID\\\":\\\"316dc856-891a-4700-afe9-6ff306b27388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:38Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:38 crc kubenswrapper[4750]: E0309 18:28:38.974781 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.373364 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.373399 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.373687 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:39 crc kubenswrapper[4750]: E0309 18:28:39.373960 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.374099 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:39 crc kubenswrapper[4750]: E0309 18:28:39.374228 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:39 crc kubenswrapper[4750]: E0309 18:28:39.374334 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:39 crc kubenswrapper[4750]: E0309 18:28:39.374433 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.394050 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a32a985-7c9a-4737-86db-d46edf0d0565\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95qv2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wl7w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.423066 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80b0e746-0802-4112-b77b-4db85dc85a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-09T18:28:16Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0309 18:28:16.291742 7363 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0309 18:28:16.291751 7363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0309 18:28:16.291766 7363 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0309 18:28:16.291771 7363 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0309 18:28:16.291783 7363 handler.go:208] Removed *v1.Node event handler 7\\\\nI0309 18:28:16.291798 7363 handler.go:208] Removed *v1.Node event handler 2\\\\nI0309 18:28:16.291840 7363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0309 18:28:16.291846 7363 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0309 18:28:16.291869 7363 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0309 18:28:16.291878 7363 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0309 18:28:16.291880 7363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0309 18:28:16.291988 7363 factory.go:656] Stopping watch factory\\\\nI0309 18:28:16.292060 7363 ovnkube.go:599] Stopped ovnkube\\\\nI0309 18:28:16.292004 7363 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0309 18:28:16.292134 7363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0309 18:28:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T18:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T18:27:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v5n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lfwr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.439180 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4c5gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ec6607b-f406-49cd-84db-c83837e9735b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://158669bee73218d7c3a4cce949cbdef16b018dd64bd02b93d5f99494f8d76281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t9dxz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4c5gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.455819 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vf7m4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc616aab-babf-4203-823a-6d5e9d5edcd3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T18:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7418422341c18f8fbe8f1c2f5090295c64f90d33f815483e4c8c67099bdb3905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T18:27:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sw7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T18:27:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vf7m4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T18:28:39Z is after 2025-08-24T17:21:41Z" Mar 09 18:28:39 crc kubenswrapper[4750]: E0309 18:28:39.498919 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.516567 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vj6kf" podStartSLOduration=131.516547664 podStartE2EDuration="2m11.516547664s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.494265275 +0000 UTC m=+200.836737683" watchObservedRunningTime="2026-03-09 18:28:39.516547664 +0000 UTC m=+200.859020062" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.558312 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podStartSLOduration=131.558286615 podStartE2EDuration="2m11.558286615s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.540319284 +0000 UTC m=+200.882791682" watchObservedRunningTime="2026-03-09 18:28:39.558286615 +0000 UTC m=+200.900759013" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.578349 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wbb5r" podStartSLOduration=130.578322773 podStartE2EDuration="2m10.578322773s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.561733909 +0000 UTC m=+200.904206327" watchObservedRunningTime="2026-03-09 18:28:39.578322773 +0000 UTC m=+200.920795171" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.608103 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=50.608081806 podStartE2EDuration="50.608081806s" podCreationTimestamp="2026-03-09 18:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.607761257 +0000 UTC m=+200.950233665" watchObservedRunningTime="2026-03-09 18:28:39.608081806 +0000 UTC m=+200.950554204" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.623204 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.623179149 podStartE2EDuration="59.623179149s" podCreationTimestamp="2026-03-09 18:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.622825179 +0000 UTC m=+200.965297597" watchObservedRunningTime="2026-03-09 18:28:39.623179149 +0000 UTC m=+200.965651547" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.684096 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=29.684067404 podStartE2EDuration="29.684067404s" podCreationTimestamp="2026-03-09 18:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.667683955 +0000 UTC m=+201.010156373" watchObservedRunningTime="2026-03-09 18:28:39.684067404 +0000 UTC m=+201.026539802" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.700841 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=73.700818191 podStartE2EDuration="1m13.700818191s" podCreationTimestamp="2026-03-09 18:27:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.684173786 +0000 UTC m=+201.026646204" watchObservedRunningTime="2026-03-09 18:28:39.700818191 +0000 UTC m=+201.043290589" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.718169 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ckgw2" podStartSLOduration=131.718144435 podStartE2EDuration="2m11.718144435s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.718080243 +0000 UTC m=+201.060552661" watchObservedRunningTime="2026-03-09 18:28:39.718144435 +0000 UTC m=+201.060616833" Mar 09 18:28:39 crc kubenswrapper[4750]: I0309 18:28:39.752906 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=37.752881304 podStartE2EDuration="37.752881304s" podCreationTimestamp="2026-03-09 18:28:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:39.739438766 +0000 UTC m=+201.081911175" watchObservedRunningTime="2026-03-09 18:28:39.752881304 +0000 UTC m=+201.095353702" Mar 09 18:28:40 crc kubenswrapper[4750]: I0309 18:28:40.373357 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:28:40 crc kubenswrapper[4750]: E0309 18:28:40.373569 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:28:41 crc kubenswrapper[4750]: I0309 18:28:41.373406 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:41 crc kubenswrapper[4750]: I0309 18:28:41.373546 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:41 crc kubenswrapper[4750]: E0309 18:28:41.373591 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:41 crc kubenswrapper[4750]: I0309 18:28:41.373684 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:41 crc kubenswrapper[4750]: E0309 18:28:41.373850 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:41 crc kubenswrapper[4750]: I0309 18:28:41.373901 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:41 crc kubenswrapper[4750]: E0309 18:28:41.373944 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:41 crc kubenswrapper[4750]: E0309 18:28:41.374023 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:43 crc kubenswrapper[4750]: I0309 18:28:43.373191 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:43 crc kubenswrapper[4750]: I0309 18:28:43.373248 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:43 crc kubenswrapper[4750]: E0309 18:28:43.374248 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:43 crc kubenswrapper[4750]: I0309 18:28:43.373398 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:43 crc kubenswrapper[4750]: I0309 18:28:43.373299 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:43 crc kubenswrapper[4750]: E0309 18:28:43.374478 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:43 crc kubenswrapper[4750]: E0309 18:28:43.374539 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:43 crc kubenswrapper[4750]: E0309 18:28:43.374613 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:44 crc kubenswrapper[4750]: E0309 18:28:44.500967 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:45 crc kubenswrapper[4750]: I0309 18:28:45.372876 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:45 crc kubenswrapper[4750]: I0309 18:28:45.372877 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:45 crc kubenswrapper[4750]: I0309 18:28:45.373070 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:45 crc kubenswrapper[4750]: I0309 18:28:45.373167 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:45 crc kubenswrapper[4750]: E0309 18:28:45.373177 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:45 crc kubenswrapper[4750]: E0309 18:28:45.373337 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:45 crc kubenswrapper[4750]: E0309 18:28:45.373329 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:45 crc kubenswrapper[4750]: E0309 18:28:45.373407 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:47 crc kubenswrapper[4750]: I0309 18:28:47.372943 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:47 crc kubenswrapper[4750]: I0309 18:28:47.373026 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:47 crc kubenswrapper[4750]: I0309 18:28:47.373100 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:47 crc kubenswrapper[4750]: I0309 18:28:47.373670 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:47 crc kubenswrapper[4750]: E0309 18:28:47.373842 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:47 crc kubenswrapper[4750]: E0309 18:28:47.373958 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:47 crc kubenswrapper[4750]: E0309 18:28:47.374043 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:47 crc kubenswrapper[4750]: E0309 18:28:47.374137 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:47 crc kubenswrapper[4750]: I0309 18:28:47.382172 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:28:47 crc kubenswrapper[4750]: I0309 18:28:47.445236 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-4c5gf" podStartSLOduration=139.445204729 podStartE2EDuration="2m19.445204729s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:47.444497199 +0000 UTC m=+208.786969597" watchObservedRunningTime="2026-03-09 18:28:47.445204729 +0000 UTC m=+208.787677127" Mar 09 18:28:47 crc kubenswrapper[4750]: I0309 18:28:47.456439 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-vf7m4" podStartSLOduration=139.456404265 podStartE2EDuration="2m19.456404265s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:47.456392774 +0000 UTC m=+208.798865172" watchObservedRunningTime="2026-03-09 18:28:47.456404265 +0000 UTC m=+208.798876673" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.004716 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.004818 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.004831 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.004857 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.004870 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T18:28:49Z","lastTransitionTime":"2026-03-09T18:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.064927 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt"] Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.065447 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.068071 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.068736 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.069655 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.069865 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.092604 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8062d269-a076-4147-a408-c89e4e19d5e3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.092702 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8062d269-a076-4147-a408-c89e4e19d5e3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.092860 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8062d269-a076-4147-a408-c89e4e19d5e3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.093004 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8062d269-a076-4147-a408-c89e4e19d5e3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.093180 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8062d269-a076-4147-a408-c89e4e19d5e3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.194484 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8062d269-a076-4147-a408-c89e4e19d5e3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.194739 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8062d269-a076-4147-a408-c89e4e19d5e3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.194806 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8062d269-a076-4147-a408-c89e4e19d5e3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.194871 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8062d269-a076-4147-a408-c89e4e19d5e3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.194871 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8062d269-a076-4147-a408-c89e4e19d5e3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.194940 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8062d269-a076-4147-a408-c89e4e19d5e3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.195548 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8062d269-a076-4147-a408-c89e4e19d5e3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.195764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8062d269-a076-4147-a408-c89e4e19d5e3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.203827 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8062d269-a076-4147-a408-c89e4e19d5e3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.212297 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8062d269-a076-4147-a408-c89e4e19d5e3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hvldt\" (UID: \"8062d269-a076-4147-a408-c89e4e19d5e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.373362 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.373420 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.373453 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.373514 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:49 crc kubenswrapper[4750]: E0309 18:28:49.374570 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:49 crc kubenswrapper[4750]: E0309 18:28:49.374852 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:49 crc kubenswrapper[4750]: E0309 18:28:49.374995 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:49 crc kubenswrapper[4750]: E0309 18:28:49.375177 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.381533 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" Mar 09 18:28:49 crc kubenswrapper[4750]: W0309 18:28:49.400898 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8062d269_a076_4147_a408_c89e4e19d5e3.slice/crio-18de33aacd872c9fdb229c6570ae1bc75d47114ada03b64eca7d211d611cbb60 WatchSource:0}: Error finding container 18de33aacd872c9fdb229c6570ae1bc75d47114ada03b64eca7d211d611cbb60: Status 404 returned error can't find the container with id 18de33aacd872c9fdb229c6570ae1bc75d47114ada03b64eca7d211d611cbb60 Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.447130 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.459876 4750 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 09 18:28:49 crc kubenswrapper[4750]: E0309 18:28:49.502397 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.599943 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" event={"ID":"8062d269-a076-4147-a408-c89e4e19d5e3","Type":"ContainerStarted","Data":"6e645440e23add043860cf0c57d7fb021659ba751f667b45c42b114e44873885"} Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.600020 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" event={"ID":"8062d269-a076-4147-a408-c89e4e19d5e3","Type":"ContainerStarted","Data":"18de33aacd872c9fdb229c6570ae1bc75d47114ada03b64eca7d211d611cbb60"} Mar 09 18:28:49 crc kubenswrapper[4750]: I0309 18:28:49.620984 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hvldt" podStartSLOduration=141.620946231 podStartE2EDuration="2m21.620946231s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:28:49.620809006 +0000 UTC m=+210.963281444" watchObservedRunningTime="2026-03-09 18:28:49.620946231 +0000 UTC m=+210.963418669" Mar 09 18:28:51 crc kubenswrapper[4750]: I0309 18:28:51.372890 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:51 crc kubenswrapper[4750]: E0309 18:28:51.373467 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:51 crc kubenswrapper[4750]: I0309 18:28:51.372964 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:51 crc kubenswrapper[4750]: E0309 18:28:51.373603 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:51 crc kubenswrapper[4750]: I0309 18:28:51.373029 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:51 crc kubenswrapper[4750]: E0309 18:28:51.373695 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:51 crc kubenswrapper[4750]: I0309 18:28:51.373126 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:51 crc kubenswrapper[4750]: E0309 18:28:51.373779 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:52 crc kubenswrapper[4750]: I0309 18:28:52.374288 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:28:52 crc kubenswrapper[4750]: E0309 18:28:52.374491 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lfwr6_openshift-ovn-kubernetes(80b0e746-0802-4112-b77b-4db85dc85a9c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" Mar 09 18:28:53 crc kubenswrapper[4750]: I0309 18:28:53.372814 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:53 crc kubenswrapper[4750]: I0309 18:28:53.372873 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:53 crc kubenswrapper[4750]: I0309 18:28:53.372918 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:53 crc kubenswrapper[4750]: I0309 18:28:53.372974 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:53 crc kubenswrapper[4750]: E0309 18:28:53.372967 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:53 crc kubenswrapper[4750]: E0309 18:28:53.373057 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:53 crc kubenswrapper[4750]: E0309 18:28:53.373236 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:53 crc kubenswrapper[4750]: E0309 18:28:53.373287 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:54 crc kubenswrapper[4750]: E0309 18:28:54.504211 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.373373 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:55 crc kubenswrapper[4750]: E0309 18:28:55.373555 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.373408 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:55 crc kubenswrapper[4750]: E0309 18:28:55.373662 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.373373 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:55 crc kubenswrapper[4750]: E0309 18:28:55.373728 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.373808 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:55 crc kubenswrapper[4750]: E0309 18:28:55.373877 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.624341 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/1.log" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.625618 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/0.log" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.625682 4750 generic.go:334] "Generic (PLEG): container finished" podID="fa652827-92a6-4544-b1b6-cba227f8e9f4" containerID="2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c" exitCode=1 Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.625715 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerDied","Data":"2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c"} Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.625751 4750 scope.go:117] "RemoveContainer" containerID="a53c6dd5b983ef287cb87ba158390067904e4125b6afee0c0aabca42968622cd" Mar 09 18:28:55 crc kubenswrapper[4750]: I0309 18:28:55.626248 4750 scope.go:117] "RemoveContainer" containerID="2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c" Mar 09 18:28:55 crc kubenswrapper[4750]: E0309 18:28:55.626452 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vj6kf_openshift-multus(fa652827-92a6-4544-b1b6-cba227f8e9f4)\"" pod="openshift-multus/multus-vj6kf" podUID="fa652827-92a6-4544-b1b6-cba227f8e9f4" Mar 09 18:28:56 crc kubenswrapper[4750]: I0309 18:28:56.630601 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/1.log" Mar 09 18:28:57 crc kubenswrapper[4750]: I0309 18:28:57.373025 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:57 crc kubenswrapper[4750]: E0309 18:28:57.373216 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:57 crc kubenswrapper[4750]: I0309 18:28:57.373430 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:57 crc kubenswrapper[4750]: I0309 18:28:57.373475 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:57 crc kubenswrapper[4750]: E0309 18:28:57.373608 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:57 crc kubenswrapper[4750]: I0309 18:28:57.373829 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:57 crc kubenswrapper[4750]: E0309 18:28:57.373878 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:57 crc kubenswrapper[4750]: E0309 18:28:57.374183 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:59 crc kubenswrapper[4750]: I0309 18:28:59.372964 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:28:59 crc kubenswrapper[4750]: I0309 18:28:59.373050 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:28:59 crc kubenswrapper[4750]: I0309 18:28:59.373046 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:28:59 crc kubenswrapper[4750]: I0309 18:28:59.373070 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:28:59 crc kubenswrapper[4750]: E0309 18:28:59.374224 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:28:59 crc kubenswrapper[4750]: E0309 18:28:59.374316 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:28:59 crc kubenswrapper[4750]: E0309 18:28:59.374422 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:28:59 crc kubenswrapper[4750]: E0309 18:28:59.374595 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:28:59 crc kubenswrapper[4750]: E0309 18:28:59.504677 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:29:01 crc kubenswrapper[4750]: I0309 18:29:01.373069 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:01 crc kubenswrapper[4750]: I0309 18:29:01.373100 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:01 crc kubenswrapper[4750]: I0309 18:29:01.373110 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:01 crc kubenswrapper[4750]: I0309 18:29:01.373154 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:01 crc kubenswrapper[4750]: E0309 18:29:01.373227 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:29:01 crc kubenswrapper[4750]: E0309 18:29:01.373325 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:29:01 crc kubenswrapper[4750]: E0309 18:29:01.373452 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:29:01 crc kubenswrapper[4750]: E0309 18:29:01.373548 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:29:03 crc kubenswrapper[4750]: I0309 18:29:03.373125 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:03 crc kubenswrapper[4750]: I0309 18:29:03.373208 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:03 crc kubenswrapper[4750]: E0309 18:29:03.373327 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:29:03 crc kubenswrapper[4750]: I0309 18:29:03.373363 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:03 crc kubenswrapper[4750]: E0309 18:29:03.373501 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:29:03 crc kubenswrapper[4750]: E0309 18:29:03.373658 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:29:03 crc kubenswrapper[4750]: I0309 18:29:03.375526 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:03 crc kubenswrapper[4750]: E0309 18:29:03.376092 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:29:04 crc kubenswrapper[4750]: E0309 18:29:04.506120 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.372976 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.373069 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.372998 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.373070 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:05 crc kubenswrapper[4750]: E0309 18:29:05.373185 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:29:05 crc kubenswrapper[4750]: E0309 18:29:05.373278 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:29:05 crc kubenswrapper[4750]: E0309 18:29:05.373389 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:29:05 crc kubenswrapper[4750]: E0309 18:29:05.373444 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.374303 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.664102 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/3.log" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.667415 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerStarted","Data":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.667899 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:29:05 crc kubenswrapper[4750]: I0309 18:29:05.735184 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podStartSLOduration=157.735165816 podStartE2EDuration="2m37.735165816s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:05.733243113 +0000 UTC m=+227.075715511" watchObservedRunningTime="2026-03-09 18:29:05.735165816 +0000 UTC m=+227.077638214" Mar 09 18:29:06 crc kubenswrapper[4750]: I0309 18:29:06.373145 4750 scope.go:117] "RemoveContainer" containerID="2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c" Mar 09 18:29:06 crc kubenswrapper[4750]: I0309 18:29:06.389999 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-wl7w6"] Mar 09 18:29:06 crc kubenswrapper[4750]: I0309 18:29:06.390265 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:06 crc kubenswrapper[4750]: E0309 18:29:06.390401 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:29:06 crc kubenswrapper[4750]: I0309 18:29:06.671788 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/1.log" Mar 09 18:29:06 crc kubenswrapper[4750]: I0309 18:29:06.671908 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerStarted","Data":"22b0048bade8f6e58ad502d87ccb4eefd6bb1e25e79954adaafc0157a254f540"} Mar 09 18:29:07 crc kubenswrapper[4750]: I0309 18:29:07.372602 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:07 crc kubenswrapper[4750]: I0309 18:29:07.372602 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:07 crc kubenswrapper[4750]: E0309 18:29:07.373049 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:29:07 crc kubenswrapper[4750]: I0309 18:29:07.372624 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:07 crc kubenswrapper[4750]: E0309 18:29:07.373164 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:29:07 crc kubenswrapper[4750]: E0309 18:29:07.373205 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:29:08 crc kubenswrapper[4750]: I0309 18:29:08.373160 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:08 crc kubenswrapper[4750]: E0309 18:29:08.373909 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wl7w6" podUID="6a32a985-7c9a-4737-86db-d46edf0d0565" Mar 09 18:29:09 crc kubenswrapper[4750]: I0309 18:29:09.372826 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:09 crc kubenswrapper[4750]: I0309 18:29:09.375717 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:09 crc kubenswrapper[4750]: E0309 18:29:09.375716 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 18:29:09 crc kubenswrapper[4750]: I0309 18:29:09.375789 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:09 crc kubenswrapper[4750]: E0309 18:29:09.376352 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 18:29:09 crc kubenswrapper[4750]: E0309 18:29:09.376674 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 18:29:10 crc kubenswrapper[4750]: I0309 18:29:10.372317 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:10 crc kubenswrapper[4750]: I0309 18:29:10.374970 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 09 18:29:10 crc kubenswrapper[4750]: I0309 18:29:10.375340 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 09 18:29:11 crc kubenswrapper[4750]: I0309 18:29:11.373204 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:11 crc kubenswrapper[4750]: I0309 18:29:11.373276 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:11 crc kubenswrapper[4750]: I0309 18:29:11.373222 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:11 crc kubenswrapper[4750]: I0309 18:29:11.376302 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 09 18:29:11 crc kubenswrapper[4750]: I0309 18:29:11.376371 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 09 18:29:11 crc kubenswrapper[4750]: I0309 18:29:11.376302 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 09 18:29:11 crc kubenswrapper[4750]: I0309 18:29:11.378089 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.665131 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.699666 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pctzl"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.700227 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.701342 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9xsqt"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.701614 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.701998 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5c9r"] Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.702475 4750 reflector.go:561] object-"openshift-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.702529 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.702569 4750 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.702586 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.702526 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.703551 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.704288 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.705545 4750 reflector.go:561] object-"openshift-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.705576 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.705703 4750 reflector.go:561] object-"openshift-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.705718 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.705938 4750 reflector.go:561] object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff": failed to list *v1.Secret: secrets "openshift-apiserver-sa-dockercfg-djjff" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.705963 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-djjff\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-sa-dockercfg-djjff\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706010 4750 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706025 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706065 4750 reflector.go:561] object-"openshift-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706078 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706118 4750 reflector.go:561] object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c": failed to list *v1.Secret: secrets "openshift-controller-manager-sa-dockercfg-msq4c" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706131 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-msq4c\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-sa-dockercfg-msq4c\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706169 4750 reflector.go:561] object-"openshift-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706182 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706221 4750 reflector.go:561] object-"openshift-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706235 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706299 4750 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706318 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706447 4750 reflector.go:561] object-"openshift-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706468 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706515 4750 reflector.go:561] object-"openshift-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706530 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706563 4750 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706575 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706644 4750 reflector.go:561] object-"openshift-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706660 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706854 4750 reflector.go:561] object-"openshift-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706878 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.706966 4750 reflector.go:561] object-"openshift-controller-manager"/"openshift-global-ca": failed to list *v1.ConfigMap: configmaps "openshift-global-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.706982 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-global-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-global-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707036 4750 reflector.go:561] object-"openshift-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707051 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707147 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707171 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707243 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707264 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707302 4750 reflector.go:561] object-"openshift-machine-api"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707319 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707508 4750 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-images": failed to list *v1.ConfigMap: configmaps "machine-api-operator-images" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707526 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-images\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-api-operator-images\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707652 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707667 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707699 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707709 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707789 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707800 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707829 4750 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707838 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707850 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707860 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.707941 4750 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.707953 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.708049 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.708061 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.708211 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.708417 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.708438 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.708471 4750 reflector.go:561] object-"openshift-oauth-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.708480 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: W0309 18:29:19.709230 4750 reflector.go:561] object-"openshift-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 09 18:29:19 crc kubenswrapper[4750]: E0309 18:29:19.709257 4750 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.713407 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jn8bm"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.714140 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.714727 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-bglg2"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.715049 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.715158 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.715212 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.715362 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.715700 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.718665 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lll4q"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.719890 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.719923 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.720607 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.720682 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.721452 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.721532 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-scg4c"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.721908 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-scg4c" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.726701 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8wcqm"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.747607 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.753778 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754126 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754173 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754186 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754130 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754525 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754608 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754657 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754800 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.754834 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.755085 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.755299 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.755392 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.755402 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.755982 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.755458 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.755529 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.756475 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757122 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757649 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757732 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757810 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757873 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757952 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758042 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757878 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758094 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757812 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758170 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758226 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758314 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758354 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758367 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758355 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758500 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.757815 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758583 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.758965 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.759378 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.759842 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.760155 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-54t4s"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.760424 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.760490 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.760529 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.761388 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.761833 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.764140 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.764608 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.765005 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qmbbg"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.765100 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.765526 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.765673 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.765988 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cwxb8"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.767493 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.767678 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.768171 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jcrcb"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.768740 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.768896 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.769011 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.769134 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.769267 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.769986 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.770200 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.770341 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.770361 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.770491 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.771043 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.771054 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.771145 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.771154 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.771159 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.771270 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.771603 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.772382 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.772520 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.772725 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.773165 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.773252 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.773270 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.773362 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.774605 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.774728 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.775271 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.775455 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.775773 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c9j6s"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.776156 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.776422 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.776693 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.776742 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.782821 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.782958 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.783576 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.783811 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.804112 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.808238 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.810707 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.825185 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.825610 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.825664 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.825614 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.826208 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.826390 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.827812 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-j5scj"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.828367 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.828528 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.829618 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.830381 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.831882 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-prpzs"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.832456 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.834775 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2dgl2"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.835408 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.836696 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bglg2"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.837238 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pctzl"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.839816 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.841167 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.846087 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551348-g9vjb"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.847152 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.847720 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.848297 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.848608 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.849657 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.853660 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854498 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-client\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854533 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r6kx\" (UniqueName: \"kubernetes.io/projected/6a5141b3-ab9a-4158-aef5-890ce185c45d-kube-api-access-7r6kx\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854559 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-console-config\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854579 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmhlh\" (UniqueName: \"kubernetes.io/projected/027f6e21-69f1-4111-85ec-f4f7486da3a5-kube-api-access-hmhlh\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854597 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-serving-cert\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854613 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-trusted-ca-bundle\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854666 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-oauth-serving-cert\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854686 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854706 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854755 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-service-ca-bundle\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854773 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhxmn\" (UniqueName: \"kubernetes.io/projected/863600af-6300-46d6-875a-85b0622c8f50-kube-api-access-jhxmn\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854790 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hc7q\" (UniqueName: \"kubernetes.io/projected/3789c695-4f63-465e-9ccb-90ee29189e38-kube-api-access-2hc7q\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854806 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854824 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-metrics-certs\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854840 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-serving-cert\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854856 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-service-ca\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854871 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7408730e-22f0-4763-b433-418ae7a4e6da-trusted-ca\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854893 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/078d41c4-0ab6-4c47-9179-c332c66882c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854912 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-config\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854928 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3484441e-7900-421d-971a-14581e0d31ff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-djq6g\" (UID: \"3484441e-7900-421d-971a-14581e0d31ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854948 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a5141b3-ab9a-4158-aef5-890ce185c45d-serving-cert\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854965 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854981 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.854996 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2m6k\" (UniqueName: \"kubernetes.io/projected/bd3871c8-50bf-434d-abb9-a457e15fba76-kube-api-access-d2m6k\") pod \"dns-operator-744455d44c-8wcqm\" (UID: \"bd3871c8-50bf-434d-abb9-a457e15fba76\") " pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855012 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-oauth-config\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855033 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/078d41c4-0ab6-4c47-9179-c332c66882c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855055 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57xrr\" (UniqueName: \"kubernetes.io/projected/43705153-a050-4d95-ba46-792661ace7ae-kube-api-access-57xrr\") pod \"downloads-7954f5f757-scg4c\" (UID: \"43705153-a050-4d95-ba46-792661ace7ae\") " pod="openshift-console/downloads-7954f5f757-scg4c" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855104 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855128 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855150 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/027f6e21-69f1-4111-85ec-f4f7486da3a5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855174 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2beaf3c7-025c-4bf7-b540-e0701c881860-machine-approver-tls\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855196 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8865d33-f955-449d-bba5-9c42abf0e7e0-serving-cert\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855219 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c91feb-2d90-44c8-a08f-d898420b1a96-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855249 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855274 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f131dc29-3928-4ef8-a3e6-d59942a59ffd-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855296 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7408730e-22f0-4763-b433-418ae7a4e6da-serving-cert\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855316 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15c91feb-2d90-44c8-a08f-d898420b1a96-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855337 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-ca\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855358 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-default-certificate\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855389 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwndd\" (UniqueName: \"kubernetes.io/projected/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-kube-api-access-hwndd\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855413 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855448 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f131dc29-3928-4ef8-a3e6-d59942a59ffd-serving-cert\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855467 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8978\" (UniqueName: \"kubernetes.io/projected/3484441e-7900-421d-971a-14581e0d31ff-kube-api-access-v8978\") pod \"cluster-samples-operator-665b6dd947-djq6g\" (UID: \"3484441e-7900-421d-971a-14581e0d31ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855485 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/078d41c4-0ab6-4c47-9179-c332c66882c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855522 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2beaf3c7-025c-4bf7-b540-e0701c881860-auth-proxy-config\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855543 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm5wr\" (UniqueName: \"kubernetes.io/projected/2beaf3c7-025c-4bf7-b540-e0701c881860-kube-api-access-tm5wr\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-client-ca\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855586 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cfb68d8f-61a3-40d7-8e89-44084f690dc2-node-pullsecrets\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855608 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2beaf3c7-025c-4bf7-b540-e0701c881860-config\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855648 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3789c695-4f63-465e-9ccb-90ee29189e38-service-ca-bundle\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855673 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855900 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssrwl\" (UniqueName: \"kubernetes.io/projected/7408730e-22f0-4763-b433-418ae7a4e6da-kube-api-access-ssrwl\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.855998 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbqwk\" (UniqueName: \"kubernetes.io/projected/cfb68d8f-61a3-40d7-8e89-44084f690dc2-kube-api-access-gbqwk\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856050 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-client\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856139 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r2jk\" (UniqueName: \"kubernetes.io/projected/15c91feb-2d90-44c8-a08f-d898420b1a96-kube-api-access-2r2jk\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856193 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-config\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856218 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvwh9\" (UniqueName: \"kubernetes.io/projected/b8865d33-f955-449d-bba5-9c42abf0e7e0-kube-api-access-jvwh9\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856236 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7408730e-22f0-4763-b433-418ae7a4e6da-config\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856286 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-stats-auth\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856308 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856367 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856392 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bd3871c8-50bf-434d-abb9-a457e15fba76-metrics-tls\") pod \"dns-operator-744455d44c-8wcqm\" (UID: \"bd3871c8-50bf-434d-abb9-a457e15fba76\") " pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856430 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit-dir\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856456 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856479 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzv25\" (UniqueName: \"kubernetes.io/projected/8d09c46d-08c7-436b-9427-49b3c9b5f875-kube-api-access-kzv25\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856513 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2fml\" (UniqueName: \"kubernetes.io/projected/078d41c4-0ab6-4c47-9179-c332c66882c5-kube-api-access-k2fml\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856531 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-config\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856577 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2jtf\" (UniqueName: \"kubernetes.io/projected/06164218-c618-4751-8f8c-d21b71f7bda5-kube-api-access-v2jtf\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856602 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856711 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d09c46d-08c7-436b-9427-49b3c9b5f875-serving-cert\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856751 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pr47\" (UniqueName: \"kubernetes.io/projected/f131dc29-3928-4ef8-a3e6-d59942a59ffd-kube-api-access-4pr47\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856791 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.856830 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-service-ca\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.863943 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.865168 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.866124 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.866479 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jn8bm"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.866503 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.866515 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lll4q"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.866526 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.866538 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-qjrwx"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.866838 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.868815 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.871736 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9xsqt"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.871768 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8wcqm"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.871779 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.871888 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.875267 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.875343 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.878839 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.878902 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.878918 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.881130 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.881371 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-prpzs"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.882858 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.883596 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.883847 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.885030 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.886636 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-scg4c"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.890005 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.892732 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5c9r"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.896440 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4w7j2"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.903078 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.906059 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.906195 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5pntt"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.923807 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c9j6s"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.923863 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.923875 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-j5scj"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.923942 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2dgl2"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.928834 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.929059 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.930595 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-54t4s"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.933188 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qjrwx"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.935510 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qmbbg"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.937206 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jcrcb"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.940401 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4w7j2"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.942471 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.942791 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5pntt"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.945845 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.946526 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.947750 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.949162 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.950465 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551348-g9vjb"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.952570 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.954352 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959261 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959302 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f131dc29-3928-4ef8-a3e6-d59942a59ffd-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959336 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7408730e-22f0-4763-b433-418ae7a4e6da-serving-cert\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959364 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15c91feb-2d90-44c8-a08f-d898420b1a96-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959387 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-ca\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959414 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-default-certificate\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959448 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwndd\" (UniqueName: \"kubernetes.io/projected/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-kube-api-access-hwndd\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959469 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959507 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f131dc29-3928-4ef8-a3e6-d59942a59ffd-serving-cert\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959526 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/078d41c4-0ab6-4c47-9179-c332c66882c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959548 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2beaf3c7-025c-4bf7-b540-e0701c881860-auth-proxy-config\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959565 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm5wr\" (UniqueName: \"kubernetes.io/projected/2beaf3c7-025c-4bf7-b540-e0701c881860-kube-api-access-tm5wr\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959589 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8978\" (UniqueName: \"kubernetes.io/projected/3484441e-7900-421d-971a-14581e0d31ff-kube-api-access-v8978\") pod \"cluster-samples-operator-665b6dd947-djq6g\" (UID: \"3484441e-7900-421d-971a-14581e0d31ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959613 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-client-ca\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959650 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cfb68d8f-61a3-40d7-8e89-44084f690dc2-node-pullsecrets\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959670 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2beaf3c7-025c-4bf7-b540-e0701c881860-config\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959693 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3789c695-4f63-465e-9ccb-90ee29189e38-service-ca-bundle\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959714 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959736 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssrwl\" (UniqueName: \"kubernetes.io/projected/7408730e-22f0-4763-b433-418ae7a4e6da-kube-api-access-ssrwl\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959760 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbqwk\" (UniqueName: \"kubernetes.io/projected/cfb68d8f-61a3-40d7-8e89-44084f690dc2-kube-api-access-gbqwk\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959783 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-client\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959809 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r2jk\" (UniqueName: \"kubernetes.io/projected/15c91feb-2d90-44c8-a08f-d898420b1a96-kube-api-access-2r2jk\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959830 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-config\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959867 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvwh9\" (UniqueName: \"kubernetes.io/projected/b8865d33-f955-449d-bba5-9c42abf0e7e0-kube-api-access-jvwh9\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959888 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7408730e-22f0-4763-b433-418ae7a4e6da-config\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959909 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-stats-auth\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959930 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959956 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.959983 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bd3871c8-50bf-434d-abb9-a457e15fba76-metrics-tls\") pod \"dns-operator-744455d44c-8wcqm\" (UID: \"bd3871c8-50bf-434d-abb9-a457e15fba76\") " pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960008 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit-dir\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960026 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzv25\" (UniqueName: \"kubernetes.io/projected/8d09c46d-08c7-436b-9427-49b3c9b5f875-kube-api-access-kzv25\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960053 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2fml\" (UniqueName: \"kubernetes.io/projected/078d41c4-0ab6-4c47-9179-c332c66882c5-kube-api-access-k2fml\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960077 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-config\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960097 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960117 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2jtf\" (UniqueName: \"kubernetes.io/projected/06164218-c618-4751-8f8c-d21b71f7bda5-kube-api-access-v2jtf\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960145 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960175 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d09c46d-08c7-436b-9427-49b3c9b5f875-serving-cert\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960197 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pr47\" (UniqueName: \"kubernetes.io/projected/f131dc29-3928-4ef8-a3e6-d59942a59ffd-kube-api-access-4pr47\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960220 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960247 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-service-ca\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960273 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-client\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960294 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r6kx\" (UniqueName: \"kubernetes.io/projected/6a5141b3-ab9a-4158-aef5-890ce185c45d-kube-api-access-7r6kx\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960315 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-console-config\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmhlh\" (UniqueName: \"kubernetes.io/projected/027f6e21-69f1-4111-85ec-f4f7486da3a5-kube-api-access-hmhlh\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960370 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-serving-cert\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960396 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-trusted-ca-bundle\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960421 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-oauth-serving-cert\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960446 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960469 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960517 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-service-ca-bundle\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960543 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhxmn\" (UniqueName: \"kubernetes.io/projected/863600af-6300-46d6-875a-85b0622c8f50-kube-api-access-jhxmn\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960580 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hc7q\" (UniqueName: \"kubernetes.io/projected/3789c695-4f63-465e-9ccb-90ee29189e38-kube-api-access-2hc7q\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960604 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960648 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-metrics-certs\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960676 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-serving-cert\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960703 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-service-ca\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960728 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7408730e-22f0-4763-b433-418ae7a4e6da-trusted-ca\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960774 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/078d41c4-0ab6-4c47-9179-c332c66882c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960802 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-config\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960847 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3484441e-7900-421d-971a-14581e0d31ff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-djq6g\" (UID: \"3484441e-7900-421d-971a-14581e0d31ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960873 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a5141b3-ab9a-4158-aef5-890ce185c45d-serving-cert\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960900 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960942 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2m6k\" (UniqueName: \"kubernetes.io/projected/bd3871c8-50bf-434d-abb9-a457e15fba76-kube-api-access-d2m6k\") pod \"dns-operator-744455d44c-8wcqm\" (UID: \"bd3871c8-50bf-434d-abb9-a457e15fba76\") " pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960963 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-oauth-config\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.960980 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/078d41c4-0ab6-4c47-9179-c332c66882c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.961000 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57xrr\" (UniqueName: \"kubernetes.io/projected/43705153-a050-4d95-ba46-792661ace7ae-kube-api-access-57xrr\") pod \"downloads-7954f5f757-scg4c\" (UID: \"43705153-a050-4d95-ba46-792661ace7ae\") " pod="openshift-console/downloads-7954f5f757-scg4c" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.961016 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.961033 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.961057 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/027f6e21-69f1-4111-85ec-f4f7486da3a5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.961084 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2beaf3c7-025c-4bf7-b540-e0701c881860-machine-approver-tls\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.961112 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8865d33-f955-449d-bba5-9c42abf0e7e0-serving-cert\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.961133 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c91feb-2d90-44c8-a08f-d898420b1a96-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.965504 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f131dc29-3928-4ef8-a3e6-d59942a59ffd-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.965569 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c91feb-2d90-44c8-a08f-d898420b1a96-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.966474 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cfb68d8f-61a3-40d7-8e89-44084f690dc2-node-pullsecrets\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.966793 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-config\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.966964 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2beaf3c7-025c-4bf7-b540-e0701c881860-config\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.967346 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-console-config\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.967829 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2beaf3c7-025c-4bf7-b540-e0701c881860-auth-proxy-config\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.968411 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.969126 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-oauth-serving-cert\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.969591 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.969651 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-p7b2r"] Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.969832 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-trusted-ca-bundle\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.969894 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-service-ca-bundle\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.970423 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.970667 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.970740 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.971193 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.971259 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-client\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.971948 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-service-ca\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.972266 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7408730e-22f0-4763-b433-418ae7a4e6da-config\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.972473 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit-dir\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.973668 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/078d41c4-0ab6-4c47-9179-c332c66882c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.974991 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a5141b3-ab9a-4158-aef5-890ce185c45d-etcd-ca\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.979701 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d09c46d-08c7-436b-9427-49b3c9b5f875-config\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.980835 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7408730e-22f0-4763-b433-418ae7a4e6da-trusted-ca\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.981532 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-oauth-config\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.982260 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2beaf3c7-025c-4bf7-b540-e0701c881860-machine-approver-tls\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.982605 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-service-ca\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.983365 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d09c46d-08c7-436b-9427-49b3c9b5f875-serving-cert\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.983831 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-serving-cert\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.986285 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f131dc29-3928-4ef8-a3e6-d59942a59ffd-serving-cert\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.987545 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/078d41c4-0ab6-4c47-9179-c332c66882c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.988308 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.988503 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a5141b3-ab9a-4158-aef5-890ce185c45d-serving-cert\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.988521 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.989020 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bd3871c8-50bf-434d-abb9-a457e15fba76-metrics-tls\") pod \"dns-operator-744455d44c-8wcqm\" (UID: \"bd3871c8-50bf-434d-abb9-a457e15fba76\") " pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.991035 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15c91feb-2d90-44c8-a08f-d898420b1a96-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.991409 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7408730e-22f0-4763-b433-418ae7a4e6da-serving-cert\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.991525 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3484441e-7900-421d-971a-14581e0d31ff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-djq6g\" (UID: \"3484441e-7900-421d-971a-14581e0d31ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.991997 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-client-ca\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.992363 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-config\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.994187 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8865d33-f955-449d-bba5-9c42abf0e7e0-serving-cert\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:19 crc kubenswrapper[4750]: I0309 18:29:19.998556 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-default-certificate\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.002324 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.008510 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-stats-auth\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.023112 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.034477 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3789c695-4f63-465e-9ccb-90ee29189e38-metrics-certs\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.043685 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.048062 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3789c695-4f63-465e-9ccb-90ee29189e38-service-ca-bundle\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.063543 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.108807 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.122528 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.142894 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.163015 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.184758 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.216373 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.227833 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.243047 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.263714 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.282360 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.303170 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.322780 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.341948 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.361985 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.382615 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.403055 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.422872 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.443037 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.469042 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.482713 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.502510 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.522585 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.542071 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.563426 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.582505 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.603926 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.622836 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.643107 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.683321 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.702574 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.723078 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.743247 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.762323 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.783165 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.802994 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.822743 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.841810 4750 request.go:700] Waited for 1.015740205s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-operator-dockercfg-2bh8d&limit=500&resourceVersion=0 Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.845385 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.863719 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.883910 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.903296 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.923620 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.942779 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.962999 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.966160 4750 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.966260 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-client podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.466231234 +0000 UTC m=+242.808703632 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-client") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.967302 4750 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.967358 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config podName:027f6e21-69f1-4111-85ec-f4f7486da3a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.467347304 +0000 UTC m=+242.809819702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config") pod "machine-api-operator-5694c8668f-p5c9r" (UID: "027f6e21-69f1-4111-85ec-f4f7486da3a5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.968645 4750 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.968702 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-serving-cert podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.468689211 +0000 UTC m=+242.811161609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-serving-cert") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.969935 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.969981 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.469969815 +0000 UTC m=+242.812442213 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.969983 4750 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.970188 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images podName:027f6e21-69f1-4111-85ec-f4f7486da3a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.470142271 +0000 UTC m=+242.812614859 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images") pod "machine-api-operator-5694c8668f-p5c9r" (UID: "027f6e21-69f1-4111-85ec-f4f7486da3a5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.970015 4750 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.970313 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.470294655 +0000 UTC m=+242.812767293 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.972421 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.972539 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.472508775 +0000 UTC m=+242.814981183 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.972826 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.972875 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.472865474 +0000 UTC m=+242.815337872 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.974181 4750 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.974261 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/027f6e21-69f1-4111-85ec-f4f7486da3a5-machine-api-operator-tls podName:027f6e21-69f1-4111-85ec-f4f7486da3a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.474239292 +0000 UTC m=+242.816711710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/027f6e21-69f1-4111-85ec-f4f7486da3a5-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-p5c9r" (UID: "027f6e21-69f1-4111-85ec-f4f7486da3a5") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.974307 4750 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.974367 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert podName:06164218-c618-4751-8f8c-d21b71f7bda5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.474348526 +0000 UTC m=+242.816821094 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert") pod "controller-manager-879f6c89f-9xsqt" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.979282 4750 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.979297 4750 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.979338 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config podName:06164218-c618-4751-8f8c-d21b71f7bda5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.479323961 +0000 UTC m=+242.821796349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config") pod "controller-manager-879f6c89f-9xsqt" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.979369 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles podName:06164218-c618-4751-8f8c-d21b71f7bda5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.479347512 +0000 UTC m=+242.821819950 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles") pod "controller-manager-879f6c89f-9xsqt" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.981099 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.981179 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.481157241 +0000 UTC m=+242.823629679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.982285 4750 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: E0309 18:29:20.982355 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca podName:06164218-c618-4751-8f8c-d21b71f7bda5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:21.482338713 +0000 UTC m=+242.824811111 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca") pod "controller-manager-879f6c89f-9xsqt" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:20 crc kubenswrapper[4750]: I0309 18:29:20.982712 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.002754 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.023436 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.043109 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.064434 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.083054 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.103396 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.123318 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.143123 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.162319 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.183356 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.203473 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.223193 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.244256 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.263835 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.284858 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.303045 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.323383 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.344121 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.362596 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.383142 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.404026 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.423266 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.445143 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.463216 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483351 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483741 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483817 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483839 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483880 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483905 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483924 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-client\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483954 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-serving-cert\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.483978 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.484002 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.484128 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.484153 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.484202 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.484220 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.484372 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/027f6e21-69f1-4111-85ec-f4f7486da3a5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.503691 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.523917 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.543819 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.562760 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.583094 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.603820 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.622978 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.643742 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.663331 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.683875 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.702656 4750 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.724461 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.743696 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.743783 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.759953 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/078d41c4-0ab6-4c47-9179-c332c66882c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.822162 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm5wr\" (UniqueName: \"kubernetes.io/projected/2beaf3c7-025c-4bf7-b540-e0701c881860-kube-api-access-tm5wr\") pod \"machine-approver-56656f9798-fdg95\" (UID: \"2beaf3c7-025c-4bf7-b540-e0701c881860\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.825967 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8978\" (UniqueName: \"kubernetes.io/projected/3484441e-7900-421d-971a-14581e0d31ff-kube-api-access-v8978\") pod \"cluster-samples-operator-665b6dd947-djq6g\" (UID: \"3484441e-7900-421d-971a-14581e0d31ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.826069 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.839711 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssrwl\" (UniqueName: \"kubernetes.io/projected/7408730e-22f0-4763-b433-418ae7a4e6da-kube-api-access-ssrwl\") pod \"console-operator-58897d9998-jn8bm\" (UID: \"7408730e-22f0-4763-b433-418ae7a4e6da\") " pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.840297 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r6kx\" (UniqueName: \"kubernetes.io/projected/6a5141b3-ab9a-4158-aef5-890ce185c45d-kube-api-access-7r6kx\") pod \"etcd-operator-b45778765-54t4s\" (UID: \"6a5141b3-ab9a-4158-aef5-890ce185c45d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.861198 4750 request.go:700] Waited for 1.893710614s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/serviceaccounts/machine-api-operator/token Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.862465 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.899385 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwndd\" (UniqueName: \"kubernetes.io/projected/bb7ec7d3-ebab-4f3d-b14d-159046f5b354-kube-api-access-hwndd\") pod \"openshift-apiserver-operator-796bbdcf4f-5w4wg\" (UID: \"bb7ec7d3-ebab-4f3d-b14d-159046f5b354\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.940218 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhxmn\" (UniqueName: \"kubernetes.io/projected/863600af-6300-46d6-875a-85b0622c8f50-kube-api-access-jhxmn\") pod \"console-f9d7485db-bglg2\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.956042 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.962937 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hc7q\" (UniqueName: \"kubernetes.io/projected/3789c695-4f63-465e-9ccb-90ee29189e38-kube-api-access-2hc7q\") pod \"router-default-5444994796-cwxb8\" (UID: \"3789c695-4f63-465e-9ccb-90ee29189e38\") " pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.969398 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.980027 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.986369 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r2jk\" (UniqueName: \"kubernetes.io/projected/15c91feb-2d90-44c8-a08f-d898420b1a96-kube-api-access-2r2jk\") pod \"openshift-controller-manager-operator-756b6f6bc6-pfpff\" (UID: \"15c91feb-2d90-44c8-a08f-d898420b1a96\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.991880 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" Mar 09 18:29:21 crc kubenswrapper[4750]: I0309 18:29:21.998660 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvwh9\" (UniqueName: \"kubernetes.io/projected/b8865d33-f955-449d-bba5-9c42abf0e7e0-kube-api-access-jvwh9\") pod \"route-controller-manager-6576b87f9c-v2sfc\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.003823 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.042170 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2m6k\" (UniqueName: \"kubernetes.io/projected/bd3871c8-50bf-434d-abb9-a457e15fba76-kube-api-access-d2m6k\") pod \"dns-operator-744455d44c-8wcqm\" (UID: \"bd3871c8-50bf-434d-abb9-a457e15fba76\") " pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.045115 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.062194 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzv25\" (UniqueName: \"kubernetes.io/projected/8d09c46d-08c7-436b-9427-49b3c9b5f875-kube-api-access-kzv25\") pod \"authentication-operator-69f744f599-lll4q\" (UID: \"8d09c46d-08c7-436b-9427-49b3c9b5f875\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.064753 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.065194 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.076781 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-54t4s"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.112529 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2fml\" (UniqueName: \"kubernetes.io/projected/078d41c4-0ab6-4c47-9179-c332c66882c5-kube-api-access-k2fml\") pod \"cluster-image-registry-operator-dc59b4c8b-97lvh\" (UID: \"078d41c4-0ab6-4c47-9179-c332c66882c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.125067 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57xrr\" (UniqueName: \"kubernetes.io/projected/43705153-a050-4d95-ba46-792661ace7ae-kube-api-access-57xrr\") pod \"downloads-7954f5f757-scg4c\" (UID: \"43705153-a050-4d95-ba46-792661ace7ae\") " pod="openshift-console/downloads-7954f5f757-scg4c" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.135703 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.152463 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.153188 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.160225 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pr47\" (UniqueName: \"kubernetes.io/projected/f131dc29-3928-4ef8-a3e6-d59942a59ffd-kube-api-access-4pr47\") pod \"openshift-config-operator-7777fb866f-tdsx8\" (UID: \"f131dc29-3928-4ef8-a3e6-d59942a59ffd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.188269 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.193971 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194331 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194388 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194446 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-etcd-client\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194478 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-certificates\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194502 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/728fe455-8ea2-4cc2-afd0-981dae4d04b0-trusted-ca\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194526 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/728fe455-8ea2-4cc2-afd0-981dae4d04b0-metrics-tls\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194610 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-bound-sa-token\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194647 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/728fe455-8ea2-4cc2-afd0-981dae4d04b0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194670 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-trusted-ca\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194691 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/483c734b-8a73-4053-b68c-256672b2f5c8-audit-dir\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194765 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194820 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-audit-policies\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194846 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw882\" (UniqueName: \"kubernetes.io/projected/483c734b-8a73-4053-b68c-256672b2f5c8-kube-api-access-fw882\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194886 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsfzt\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-kube-api-access-lsfzt\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194918 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwl4\" (UniqueName: \"kubernetes.io/projected/728fe455-8ea2-4cc2-afd0-981dae4d04b0-kube-api-access-dwwl4\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194937 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-tls\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.194974 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-encryption-config\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.195044 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.195065 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.195116 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-serving-cert\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.195509 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:22.695488404 +0000 UTC m=+244.037960802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.209137 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.217761 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.222686 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.235525 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bglg2"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.243492 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/027f6e21-69f1-4111-85ec-f4f7486da3a5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.243526 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.266313 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.302010 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8wcqm"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304402 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304694 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-audit-policies\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304737 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6708629c-b5b3-4817-add5-f39c09860086-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304770 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tlh8\" (UniqueName: \"kubernetes.io/projected/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-kube-api-access-4tlh8\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304795 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9d4l\" (UniqueName: \"kubernetes.io/projected/f821a876-eab1-4298-b06c-9560fb385085-kube-api-access-f9d4l\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304827 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93426711-c51f-4f98-b0aa-e0478447375d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304860 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ecde8f-d9fa-45e5-a508-981772436b4f-secret-volume\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304880 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfrz6\" (UniqueName: \"kubernetes.io/projected/22da2745-194a-4eea-bae4-923bdbaf6cab-kube-api-access-rfrz6\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304900 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304956 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.304980 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsfzt\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-kube-api-access-lsfzt\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305036 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305064 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3174c30-7f17-4ed8-b319-2ffc759ad14d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305101 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-images\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305127 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978fa61e-9c1b-4aae-8fef-e3763342994c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305150 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5ngq\" (UniqueName: \"kubernetes.io/projected/93169b48-d875-458c-b7bd-ee2f4fd250f9-kube-api-access-z5ngq\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305177 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-encryption-config\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305215 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-profile-collector-cert\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305254 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-socket-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305285 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ecde8f-d9fa-45e5-a508-981772436b4f-config-volume\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305308 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3174c30-7f17-4ed8-b319-2ffc759ad14d-config\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305335 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/653a7146-32eb-46ad-88bf-d16dddb2192f-webhook-cert\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305356 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-csi-data-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305376 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqp4f\" (UniqueName: \"kubernetes.io/projected/49d76d7c-f89a-426a-8855-95b847669e5e-kube-api-access-cqp4f\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305396 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-proxy-tls\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305417 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvs8x\" (UniqueName: \"kubernetes.io/projected/93426711-c51f-4f98-b0aa-e0478447375d-kube-api-access-cvs8x\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305475 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3174c30-7f17-4ed8-b319-2ffc759ad14d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305559 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305583 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305604 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-registration-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305623 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08aaa5f-186f-4f67-9332-d1b218a24943-config\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305663 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305735 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/71bbf2b6-cac6-4959-9124-607c988da153-cert\") pod \"ingress-canary-4w7j2\" (UID: \"71bbf2b6-cac6-4959-9124-607c988da153\") " pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305762 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf2ml\" (UniqueName: \"kubernetes.io/projected/dac5b814-763c-42f7-a499-7a6afe1787b9-kube-api-access-qf2ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5d9d\" (UID: \"dac5b814-763c-42f7-a499-7a6afe1787b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305784 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305824 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/978fa61e-9c1b-4aae-8fef-e3763342994c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305850 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93169b48-d875-458c-b7bd-ee2f4fd250f9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305875 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305915 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnbl6\" (UniqueName: \"kubernetes.io/projected/653a7146-32eb-46ad-88bf-d16dddb2192f-kube-api-access-xnbl6\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.305939 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c08aaa5f-186f-4f67-9332-d1b218a24943-serving-cert\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306053 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306128 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-etcd-client\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306151 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/800349eb-4f6c-475e-9d56-c69c67b207a8-node-bootstrap-token\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306185 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt8qn\" (UniqueName: \"kubernetes.io/projected/71bbf2b6-cac6-4959-9124-607c988da153-kube-api-access-rt8qn\") pod \"ingress-canary-4w7j2\" (UID: \"71bbf2b6-cac6-4959-9124-607c988da153\") " pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306204 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22da2745-194a-4eea-bae4-923bdbaf6cab-config-volume\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306225 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sflfj\" (UniqueName: \"kubernetes.io/projected/90b2ac21-294a-40b6-b91e-8b189e5cca78-kube-api-access-sflfj\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306248 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/49d76d7c-f89a-426a-8855-95b847669e5e-signing-cabundle\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306270 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-audit-policies\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306307 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z64rx\" (UniqueName: \"kubernetes.io/projected/0e15e4fc-3419-4de4-b5db-6b10b1d4b962-kube-api-access-z64rx\") pod \"package-server-manager-789f6589d5-svpt8\" (UID: \"0e15e4fc-3419-4de4-b5db-6b10b1d4b962\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306326 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6708629c-b5b3-4817-add5-f39c09860086-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306351 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/332ab927-566e-4de9-bb3b-9784f123146b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-j5scj\" (UID: \"332ab927-566e-4de9-bb3b-9784f123146b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306449 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/728fe455-8ea2-4cc2-afd0-981dae4d04b0-trusted-ca\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306499 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93169b48-d875-458c-b7bd-ee2f4fd250f9-proxy-tls\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306531 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/728fe455-8ea2-4cc2-afd0-981dae4d04b0-metrics-tls\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306577 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/728fe455-8ea2-4cc2-afd0-981dae4d04b0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306600 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/483c734b-8a73-4053-b68c-256672b2f5c8-audit-dir\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306622 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306694 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306717 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22da2745-194a-4eea-bae4-923bdbaf6cab-metrics-tls\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306739 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306758 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306803 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw882\" (UniqueName: \"kubernetes.io/projected/483c734b-8a73-4053-b68c-256672b2f5c8-kube-api-access-fw882\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306826 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dac5b814-763c-42f7-a499-7a6afe1787b9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5d9d\" (UID: \"dac5b814-763c-42f7-a499-7a6afe1787b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306879 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-srv-cert\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306910 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f821a876-eab1-4298-b06c-9560fb385085-audit-dir\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.306980 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69mvv\" (UniqueName: \"kubernetes.io/projected/332ab927-566e-4de9-bb3b-9784f123146b-kube-api-access-69mvv\") pod \"multus-admission-controller-857f4d67dd-j5scj\" (UID: \"332ab927-566e-4de9-bb3b-9784f123146b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307006 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fkqp\" (UniqueName: \"kubernetes.io/projected/36c2456b-5e37-4a71-9c55-7e1bd2182512-kube-api-access-9fkqp\") pod \"migrator-59844c95c7-b84rp\" (UID: \"36c2456b-5e37-4a71-9c55-7e1bd2182512\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307032 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwl4\" (UniqueName: \"kubernetes.io/projected/728fe455-8ea2-4cc2-afd0-981dae4d04b0-kube-api-access-dwwl4\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307053 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307090 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-tls\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307139 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6708629c-b5b3-4817-add5-f39c09860086-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307181 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmd9\" (UniqueName: \"kubernetes.io/projected/d0ecde8f-d9fa-45e5-a508-981772436b4f-kube-api-access-dcmd9\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307251 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307285 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307333 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk6ng\" (UniqueName: \"kubernetes.io/projected/a08fcb79-42d8-47c2-940e-92c4ff4f2f61-kube-api-access-fk6ng\") pod \"auto-csr-approver-29551348-g9vjb\" (UID: \"a08fcb79-42d8-47c2-940e-92c4ff4f2f61\") " pod="openshift-infra/auto-csr-approver-29551348-g9vjb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307357 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdhrn\" (UniqueName: \"kubernetes.io/projected/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-kube-api-access-hdhrn\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307459 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-mountpoint-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307516 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-serving-cert\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307580 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnhg8\" (UniqueName: \"kubernetes.io/projected/800349eb-4f6c-475e-9d56-c69c67b207a8-kube-api-access-dnhg8\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.307793 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:22.807769163 +0000 UTC m=+244.150241561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.307622 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.323746 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.324187 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.332934 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/728fe455-8ea2-4cc2-afd0-981dae4d04b0-trusted-ca\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.333509 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.337349 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.343260 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.343321 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/728fe455-8ea2-4cc2-afd0-981dae4d04b0-metrics-tls\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.350685 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.351248 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/483c734b-8a73-4053-b68c-256672b2f5c8-audit-dir\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.351452 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-scg4c" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.351961 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/653a7146-32eb-46ad-88bf-d16dddb2192f-apiservice-cert\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.352010 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-plugins-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.352063 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93426711-c51f-4f98-b0aa-e0478447375d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.352082 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.352427 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/49d76d7c-f89a-426a-8855-95b847669e5e-signing-key\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.352511 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.352591 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-certificates\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.352613 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5p7g\" (UniqueName: \"kubernetes.io/projected/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-kube-api-access-w5p7g\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.354595 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-certificates\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.356586 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.357414 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-tls\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.357855 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-serving-cert\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.360646 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-encryption-config\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361438 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e15e4fc-3419-4de4-b5db-6b10b1d4b962-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-svpt8\" (UID: \"0e15e4fc-3419-4de4-b5db-6b10b1d4b962\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361488 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbbbz\" (UniqueName: \"kubernetes.io/projected/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-kube-api-access-cbbbz\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361552 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-bound-sa-token\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361575 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-srv-cert\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361625 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/653a7146-32eb-46ad-88bf-d16dddb2192f-tmpfs\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361684 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-trusted-ca\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361715 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/978fa61e-9c1b-4aae-8fef-e3763342994c-config\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.361738 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/800349eb-4f6c-475e-9d56-c69c67b207a8-certs\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.362444 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckqqr\" (UniqueName: \"kubernetes.io/projected/c08aaa5f-186f-4f67-9332-d1b218a24943-kube-api-access-ckqqr\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.362823 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.365212 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-trusted-ca\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.365331 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.390948 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jn8bm"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.392560 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.393926 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-client\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.394521 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.403483 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.403960 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.407876 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.423900 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.448963 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.463858 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.466585 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnbl6\" (UniqueName: \"kubernetes.io/projected/653a7146-32eb-46ad-88bf-d16dddb2192f-kube-api-access-xnbl6\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.466622 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c08aaa5f-186f-4f67-9332-d1b218a24943-serving-cert\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.466643 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467063 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/800349eb-4f6c-475e-9d56-c69c67b207a8-node-bootstrap-token\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467085 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt8qn\" (UniqueName: \"kubernetes.io/projected/71bbf2b6-cac6-4959-9124-607c988da153-kube-api-access-rt8qn\") pod \"ingress-canary-4w7j2\" (UID: \"71bbf2b6-cac6-4959-9124-607c988da153\") " pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467113 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22da2745-194a-4eea-bae4-923bdbaf6cab-config-volume\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467118 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467128 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sflfj\" (UniqueName: \"kubernetes.io/projected/90b2ac21-294a-40b6-b91e-8b189e5cca78-kube-api-access-sflfj\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467248 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/49d76d7c-f89a-426a-8855-95b847669e5e-signing-cabundle\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467269 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-audit-policies\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467304 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z64rx\" (UniqueName: \"kubernetes.io/projected/0e15e4fc-3419-4de4-b5db-6b10b1d4b962-kube-api-access-z64rx\") pod \"package-server-manager-789f6589d5-svpt8\" (UID: \"0e15e4fc-3419-4de4-b5db-6b10b1d4b962\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/332ab927-566e-4de9-bb3b-9784f123146b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-j5scj\" (UID: \"332ab927-566e-4de9-bb3b-9784f123146b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467371 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6708629c-b5b3-4817-add5-f39c09860086-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467430 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93169b48-d875-458c-b7bd-ee2f4fd250f9-proxy-tls\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467466 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467508 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22da2745-194a-4eea-bae4-923bdbaf6cab-metrics-tls\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467528 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467544 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467585 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dac5b814-763c-42f7-a499-7a6afe1787b9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5d9d\" (UID: \"dac5b814-763c-42f7-a499-7a6afe1787b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467609 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-srv-cert\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467653 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f821a876-eab1-4298-b06c-9560fb385085-audit-dir\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467669 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fkqp\" (UniqueName: \"kubernetes.io/projected/36c2456b-5e37-4a71-9c55-7e1bd2182512-kube-api-access-9fkqp\") pod \"migrator-59844c95c7-b84rp\" (UID: \"36c2456b-5e37-4a71-9c55-7e1bd2182512\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467688 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69mvv\" (UniqueName: \"kubernetes.io/projected/332ab927-566e-4de9-bb3b-9784f123146b-kube-api-access-69mvv\") pod \"multus-admission-controller-857f4d67dd-j5scj\" (UID: \"332ab927-566e-4de9-bb3b-9784f123146b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.467704 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.469156 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22da2745-194a-4eea-bae4-923bdbaf6cab-config-volume\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.469304 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f821a876-eab1-4298-b06c-9560fb385085-audit-dir\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.470684 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6708629c-b5b3-4817-add5-f39c09860086-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.470733 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmd9\" (UniqueName: \"kubernetes.io/projected/d0ecde8f-d9fa-45e5-a508-981772436b4f-kube-api-access-dcmd9\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.470798 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.470824 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk6ng\" (UniqueName: \"kubernetes.io/projected/a08fcb79-42d8-47c2-940e-92c4ff4f2f61-kube-api-access-fk6ng\") pod \"auto-csr-approver-29551348-g9vjb\" (UID: \"a08fcb79-42d8-47c2-940e-92c4ff4f2f61\") " pod="openshift-infra/auto-csr-approver-29551348-g9vjb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.471387 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-audit-policies\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.471654 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.471840 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.472113 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdhrn\" (UniqueName: \"kubernetes.io/projected/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-kube-api-access-hdhrn\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.472140 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-mountpoint-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.473186 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:22.973165224 +0000 UTC m=+244.315637622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.474213 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-mountpoint-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.474733 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/49d76d7c-f89a-426a-8855-95b847669e5e-signing-cabundle\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.474830 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6708629c-b5b3-4817-add5-f39c09860086-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476704 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnhg8\" (UniqueName: \"kubernetes.io/projected/800349eb-4f6c-475e-9d56-c69c67b207a8-kube-api-access-dnhg8\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476800 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-plugins-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476847 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93426711-c51f-4f98-b0aa-e0478447375d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476870 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476893 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/653a7146-32eb-46ad-88bf-d16dddb2192f-apiservice-cert\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476935 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/49d76d7c-f89a-426a-8855-95b847669e5e-signing-key\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476955 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.476989 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5p7g\" (UniqueName: \"kubernetes.io/projected/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-kube-api-access-w5p7g\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477036 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbbbz\" (UniqueName: \"kubernetes.io/projected/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-kube-api-access-cbbbz\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477059 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e15e4fc-3419-4de4-b5db-6b10b1d4b962-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-svpt8\" (UID: \"0e15e4fc-3419-4de4-b5db-6b10b1d4b962\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477099 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-srv-cert\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477136 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/653a7146-32eb-46ad-88bf-d16dddb2192f-tmpfs\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477156 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/978fa61e-9c1b-4aae-8fef-e3763342994c-config\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477191 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/800349eb-4f6c-475e-9d56-c69c67b207a8-certs\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477218 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckqqr\" (UniqueName: \"kubernetes.io/projected/c08aaa5f-186f-4f67-9332-d1b218a24943-kube-api-access-ckqqr\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477260 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6708629c-b5b3-4817-add5-f39c09860086-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477280 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tlh8\" (UniqueName: \"kubernetes.io/projected/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-kube-api-access-4tlh8\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477304 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93426711-c51f-4f98-b0aa-e0478447375d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477346 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9d4l\" (UniqueName: \"kubernetes.io/projected/f821a876-eab1-4298-b06c-9560fb385085-kube-api-access-f9d4l\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477365 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfrz6\" (UniqueName: \"kubernetes.io/projected/22da2745-194a-4eea-bae4-923bdbaf6cab-kube-api-access-rfrz6\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477382 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.477405 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ecde8f-d9fa-45e5-a508-981772436b4f-secret-volume\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.478475 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.478509 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.478553 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3174c30-7f17-4ed8-b319-2ffc759ad14d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.478684 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-images\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.478762 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5ngq\" (UniqueName: \"kubernetes.io/projected/93169b48-d875-458c-b7bd-ee2f4fd250f9-kube-api-access-z5ngq\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.478785 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978fa61e-9c1b-4aae-8fef-e3763342994c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.479561 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-profile-collector-cert\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.479592 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ecde8f-d9fa-45e5-a508-981772436b4f-config-volume\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.479650 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-socket-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.479669 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/653a7146-32eb-46ad-88bf-d16dddb2192f-webhook-cert\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.479741 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3174c30-7f17-4ed8-b319-2ffc759ad14d-config\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.479768 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqp4f\" (UniqueName: \"kubernetes.io/projected/49d76d7c-f89a-426a-8855-95b847669e5e-kube-api-access-cqp4f\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.480174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-proxy-tls\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.480199 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvs8x\" (UniqueName: \"kubernetes.io/projected/93426711-c51f-4f98-b0aa-e0478447375d-kube-api-access-cvs8x\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.481198 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93426711-c51f-4f98-b0aa-e0478447375d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.481598 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-plugins-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.483114 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.483274 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/653a7146-32eb-46ad-88bf-d16dddb2192f-tmpfs\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.483764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/978fa61e-9c1b-4aae-8fef-e3763342994c-config\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.484620 4750 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.485197 4750 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.485219 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.485871 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.485864 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-csi-data-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.485937 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3174c30-7f17-4ed8-b319-2ffc759ad14d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486280 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486316 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486371 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08aaa5f-186f-4f67-9332-d1b218a24943-config\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486394 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486416 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-registration-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486455 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/71bbf2b6-cac6-4959-9124-607c988da153-cert\") pod \"ingress-canary-4w7j2\" (UID: \"71bbf2b6-cac6-4959-9124-607c988da153\") " pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486475 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf2ml\" (UniqueName: \"kubernetes.io/projected/dac5b814-763c-42f7-a499-7a6afe1787b9-kube-api-access-qf2ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5d9d\" (UID: \"dac5b814-763c-42f7-a499-7a6afe1787b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486499 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486557 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93169b48-d875-458c-b7bd-ee2f4fd250f9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486585 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/978fa61e-9c1b-4aae-8fef-e3763342994c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.486692 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-images\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.486785 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.486960 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.487126 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config podName:06164218-c618-4751-8f8c-d21b71f7bda5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.487108165 +0000 UTC m=+244.829580563 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config") pod "controller-manager-879f6c89f-9xsqt" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.487165 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images podName:027f6e21-69f1-4111-85ec-f4f7486da3a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.487142586 +0000 UTC m=+244.829614984 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images") pod "machine-api-operator-5694c8668f-p5c9r" (UID: "027f6e21-69f1-4111-85ec-f4f7486da3a5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.487368 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.487352072 +0000 UTC m=+244.829824470 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.488841 4750 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.488973 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.488941245 +0000 UTC m=+244.831413813 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.490459 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.490435106 +0000 UTC m=+244.832907504 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.491723 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.494555 4750 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.494756 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca podName:06164218-c618-4751-8f8c-d21b71f7bda5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.494738103 +0000 UTC m=+244.837210501 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca") pod "controller-manager-879f6c89f-9xsqt" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.495196 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-registration-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.495444 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-socket-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.497785 4750 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.497825 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.497850 4750 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.498955 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/90b2ac21-294a-40b6-b91e-8b189e5cca78-csi-data-dir\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.499233 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config podName:027f6e21-69f1-4111-85ec-f4f7486da3a5 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.499213706 +0000 UTC m=+244.841686104 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config") pod "machine-api-operator-5694c8668f-p5c9r" (UID: "027f6e21-69f1-4111-85ec-f4f7486da3a5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.499275 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.499269418 +0000 UTC m=+244.841741816 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.499345 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca podName:cfb68d8f-61a3-40d7-8e89-44084f690dc2 nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.499309419 +0000 UTC m=+244.841781817 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca") pod "apiserver-76f77b778f-pctzl" (UID: "cfb68d8f-61a3-40d7-8e89-44084f690dc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.499875 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.501561 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08aaa5f-186f-4f67-9332-d1b218a24943-config\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.501608 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-profile-collector-cert\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.502091 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6708629c-b5b3-4817-add5-f39c09860086-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.502461 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.502852 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93169b48-d875-458c-b7bd-ee2f4fd250f9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.502881 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.503581 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22da2745-194a-4eea-bae4-923bdbaf6cab-metrics-tls\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.503833 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ecde8f-d9fa-45e5-a508-981772436b4f-config-volume\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.503986 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e15e4fc-3419-4de4-b5db-6b10b1d4b962-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-svpt8\" (UID: \"0e15e4fc-3419-4de4-b5db-6b10b1d4b962\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.505167 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c08aaa5f-186f-4f67-9332-d1b218a24943-serving-cert\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.507141 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/800349eb-4f6c-475e-9d56-c69c67b207a8-node-bootstrap-token\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.507434 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.507438 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/71bbf2b6-cac6-4959-9124-607c988da153-cert\") pod \"ingress-canary-4w7j2\" (UID: \"71bbf2b6-cac6-4959-9124-607c988da153\") " pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.507666 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/800349eb-4f6c-475e-9d56-c69c67b207a8-certs\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.510019 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/332ab927-566e-4de9-bb3b-9784f123146b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-j5scj\" (UID: \"332ab927-566e-4de9-bb3b-9784f123146b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.510039 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.510106 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-audit-policies\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.510349 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3174c30-7f17-4ed8-b319-2ffc759ad14d-config\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.510741 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93426711-c51f-4f98-b0aa-e0478447375d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.510988 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.511201 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.511348 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.512009 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-srv-cert\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.512017 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3174c30-7f17-4ed8-b319-2ffc759ad14d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.512836 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-srv-cert\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.513936 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-proxy-tls\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.514648 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93169b48-d875-458c-b7bd-ee2f4fd250f9-proxy-tls\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.516192 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dac5b814-763c-42f7-a499-7a6afe1787b9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5d9d\" (UID: \"dac5b814-763c-42f7-a499-7a6afe1787b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.517150 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.520053 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/978fa61e-9c1b-4aae-8fef-e3763342994c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.520113 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.520342 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.520511 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ecde8f-d9fa-45e5-a508-981772436b4f-secret-volume\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.521458 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/653a7146-32eb-46ad-88bf-d16dddb2192f-apiservice-cert\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.531612 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/653a7146-32eb-46ad-88bf-d16dddb2192f-webhook-cert\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.532522 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.537333 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh"] Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.548187 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.555110 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-serving-cert\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.569473 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.571313 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.574400 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/49d76d7c-f89a-426a-8855-95b847669e5e-signing-key\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.586379 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.588271 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.589082 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.089061992 +0000 UTC m=+244.431534390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: W0309 18:29:22.594458 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8865d33_f955_449d_bba5_9c42abf0e7e0.slice/crio-f7f19f6f861ebbc1b0f4af95b0f0db615cade4a08c1b3cdb2804bc1001887e92 WatchSource:0}: Error finding container f7f19f6f861ebbc1b0f4af95b0f0db615cade4a08c1b3cdb2804bc1001887e92: Status 404 returned error can't find the container with id f7f19f6f861ebbc1b0f4af95b0f0db615cade4a08c1b3cdb2804bc1001887e92 Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.616729 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.631468 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.641085 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.647253 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.662561 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbqwk\" (UniqueName: \"kubernetes.io/projected/cfb68d8f-61a3-40d7-8e89-44084f690dc2-kube-api-access-gbqwk\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.667563 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.680091 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2jtf\" (UniqueName: \"kubernetes.io/projected/06164218-c618-4751-8f8c-d21b71f7bda5-kube-api-access-v2jtf\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.683647 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.690652 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.691116 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.191101411 +0000 UTC m=+244.533573809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.694424 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/483c734b-8a73-4053-b68c-256672b2f5c8-etcd-client\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.703134 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.722485 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.742762 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.751875 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" event={"ID":"15c91feb-2d90-44c8-a08f-d898420b1a96","Type":"ContainerStarted","Data":"5aa0d553933a86ffde0cbf4027308b19ca69334505ffc8dd6be6e15450cb16b4"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.751924 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" event={"ID":"15c91feb-2d90-44c8-a08f-d898420b1a96","Type":"ContainerStarted","Data":"c151879059bb6b4daad638c0d6cf22b5f248931fb7f83df56824271e52582ccd"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.755809 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" event={"ID":"078d41c4-0ab6-4c47-9179-c332c66882c5","Type":"ContainerStarted","Data":"1947f3ec616c62605c3fb1ea2de37acd0a99f7c6aca97200586cfb84620a81e3"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.761510 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bglg2" event={"ID":"863600af-6300-46d6-875a-85b0622c8f50","Type":"ContainerStarted","Data":"71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.761567 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bglg2" event={"ID":"863600af-6300-46d6-875a-85b0622c8f50","Type":"ContainerStarted","Data":"8a0387233f74db99d6e02b6cbfa634ae69172f4c272ca750011d1b0cf121ca47"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.763780 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.769489 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" event={"ID":"6a5141b3-ab9a-4158-aef5-890ce185c45d","Type":"ContainerStarted","Data":"e7693d22ba282a35890d93d116c31559d30c004fc56a35d7b967d9d461ed016d"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.769573 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" event={"ID":"6a5141b3-ab9a-4158-aef5-890ce185c45d","Type":"ContainerStarted","Data":"ed6ff2cd655dde7b1d6a78133de6c21e243a05df515e092637eaa66392510393"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.772066 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cwxb8" event={"ID":"3789c695-4f63-465e-9ccb-90ee29189e38","Type":"ContainerStarted","Data":"91221e57099b387a62deb5894a10449cd84c7b648b6ca373e5af12b998a6e8d7"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.772119 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cwxb8" event={"ID":"3789c695-4f63-465e-9ccb-90ee29189e38","Type":"ContainerStarted","Data":"7cace2621264bacfa11b2898669e2e34528a61585eb6ec373e880c86b8cd2cb7"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.775620 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" event={"ID":"bd3871c8-50bf-434d-abb9-a457e15fba76","Type":"ContainerStarted","Data":"ada0650bb2e437eb2de7d00895fbff1a532c0de5ae1507869124139a56aef390"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.776466 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" event={"ID":"3484441e-7900-421d-971a-14581e0d31ff","Type":"ContainerStarted","Data":"3bab95ad75eac6069362e4fcf0e557731e5061f67319909ad49489821e716928"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.779196 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" event={"ID":"2beaf3c7-025c-4bf7-b540-e0701c881860","Type":"ContainerStarted","Data":"d4adf31728de588e37b74627ff4ff49bdbe847c97c48dffbaab8a4a4f5a60b97"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.779255 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" event={"ID":"2beaf3c7-025c-4bf7-b540-e0701c881860","Type":"ContainerStarted","Data":"971f8941690526628332b732ac32515d824e56e3e4bc3576a418ffe2a2445cd9"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.783350 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.785109 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" event={"ID":"7408730e-22f0-4763-b433-418ae7a4e6da","Type":"ContainerStarted","Data":"acc06f7dcc2895052678d2ec7358326a4f2a35929333bf28c86faf4c42fc75a1"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.785176 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" event={"ID":"7408730e-22f0-4763-b433-418ae7a4e6da","Type":"ContainerStarted","Data":"ad6f96cfcab08b75f9e1892a57463a24238381a9fcbbd93c44810d771cda0ce5"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.787364 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.787432 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" event={"ID":"b8865d33-f955-449d-bba5-9c42abf0e7e0","Type":"ContainerStarted","Data":"f7f19f6f861ebbc1b0f4af95b0f0db615cade4a08c1b3cdb2804bc1001887e92"} Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.789887 4750 patch_prober.go:28] interesting pod/console-operator-58897d9998-jn8bm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.789955 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" podUID="7408730e-22f0-4763-b433-418ae7a4e6da" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.791498 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.791620 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.291570727 +0000 UTC m=+244.634043125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.792164 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.792674 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.292665737 +0000 UTC m=+244.635138135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.795175 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/483c734b-8a73-4053-b68c-256672b2f5c8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.803123 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.808878 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmhlh\" (UniqueName: \"kubernetes.io/projected/027f6e21-69f1-4111-85ec-f4f7486da3a5-kube-api-access-hmhlh\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.823255 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.886575 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw882\" (UniqueName: \"kubernetes.io/projected/483c734b-8a73-4053-b68c-256672b2f5c8-kube-api-access-fw882\") pod \"apiserver-7bbb656c7d-j765j\" (UID: \"483c734b-8a73-4053-b68c-256672b2f5c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.893429 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.895807 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.395780096 +0000 UTC m=+244.738252524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.905972 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsfzt\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-kube-api-access-lsfzt\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.918123 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwl4\" (UniqueName: \"kubernetes.io/projected/728fe455-8ea2-4cc2-afd0-981dae4d04b0-kube-api-access-dwwl4\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.948645 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/728fe455-8ea2-4cc2-afd0-981dae4d04b0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cj8qp\" (UID: \"728fe455-8ea2-4cc2-afd0-981dae4d04b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.961133 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-bound-sa-token\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.978915 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sflfj\" (UniqueName: \"kubernetes.io/projected/90b2ac21-294a-40b6-b91e-8b189e5cca78-kube-api-access-sflfj\") pod \"csi-hostpathplugin-5pntt\" (UID: \"90b2ac21-294a-40b6-b91e-8b189e5cca78\") " pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.988193 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" Mar 09 18:29:22 crc kubenswrapper[4750]: I0309 18:29:22.996251 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:22 crc kubenswrapper[4750]: E0309 18:29:22.998124 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.498107843 +0000 UTC m=+244.840580241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.024704 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnbl6\" (UniqueName: \"kubernetes.io/projected/653a7146-32eb-46ad-88bf-d16dddb2192f-kube-api-access-xnbl6\") pod \"packageserver-d55dfcdfc-hlcdk\" (UID: \"653a7146-32eb-46ad-88bf-d16dddb2192f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.052799 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt8qn\" (UniqueName: \"kubernetes.io/projected/71bbf2b6-cac6-4959-9124-607c988da153-kube-api-access-rt8qn\") pod \"ingress-canary-4w7j2\" (UID: \"71bbf2b6-cac6-4959-9124-607c988da153\") " pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.054820 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-scg4c"] Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.060356 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lll4q"] Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.062543 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg"] Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.067127 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69mvv\" (UniqueName: \"kubernetes.io/projected/332ab927-566e-4de9-bb3b-9784f123146b-kube-api-access-69mvv\") pod \"multus-admission-controller-857f4d67dd-j5scj\" (UID: \"332ab927-566e-4de9-bb3b-9784f123146b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.073861 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.081274 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6708629c-b5b3-4817-add5-f39c09860086-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qg758\" (UID: \"6708629c-b5b3-4817-add5-f39c09860086\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.088371 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fkqp\" (UniqueName: \"kubernetes.io/projected/36c2456b-5e37-4a71-9c55-7e1bd2182512-kube-api-access-9fkqp\") pod \"migrator-59844c95c7-b84rp\" (UID: \"36c2456b-5e37-4a71-9c55-7e1bd2182512\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.098242 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.099184 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.599162665 +0000 UTC m=+244.941635063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.099527 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8"] Mar 09 18:29:23 crc kubenswrapper[4750]: W0309 18:29:23.112698 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb7ec7d3_ebab_4f3d_b14d_159046f5b354.slice/crio-49d3396f0fbc795451ea27a718eb0549c10843c5d4717ff11d63cb498d9157ba WatchSource:0}: Error finding container 49d3396f0fbc795451ea27a718eb0549c10843c5d4717ff11d63cb498d9157ba: Status 404 returned error can't find the container with id 49d3396f0fbc795451ea27a718eb0549c10843c5d4717ff11d63cb498d9157ba Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.118183 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z64rx\" (UniqueName: \"kubernetes.io/projected/0e15e4fc-3419-4de4-b5db-6b10b1d4b962-kube-api-access-z64rx\") pod \"package-server-manager-789f6589d5-svpt8\" (UID: \"0e15e4fc-3419-4de4-b5db-6b10b1d4b962\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.123415 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" Mar 09 18:29:23 crc kubenswrapper[4750]: W0309 18:29:23.129729 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf131dc29_3928_4ef8_a3e6_d59942a59ffd.slice/crio-463e304e1607e422055db4ca2498af4b9f38cc4823258f9ee7ca5288d4c56826 WatchSource:0}: Error finding container 463e304e1607e422055db4ca2498af4b9f38cc4823258f9ee7ca5288d4c56826: Status 404 returned error can't find the container with id 463e304e1607e422055db4ca2498af4b9f38cc4823258f9ee7ca5288d4c56826 Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.133274 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmd9\" (UniqueName: \"kubernetes.io/projected/d0ecde8f-d9fa-45e5-a508-981772436b4f-kube-api-access-dcmd9\") pod \"collect-profiles-29551335-l5ldr\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.138916 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.149948 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk6ng\" (UniqueName: \"kubernetes.io/projected/a08fcb79-42d8-47c2-940e-92c4ff4f2f61-kube-api-access-fk6ng\") pod \"auto-csr-approver-29551348-g9vjb\" (UID: \"a08fcb79-42d8-47c2-940e-92c4ff4f2f61\") " pod="openshift-infra/auto-csr-approver-29551348-g9vjb" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.170611 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.170992 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdhrn\" (UniqueName: \"kubernetes.io/projected/57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3-kube-api-access-hdhrn\") pod \"machine-config-operator-74547568cd-94ztz\" (UID: \"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.174816 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.195192 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.195821 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnhg8\" (UniqueName: \"kubernetes.io/projected/800349eb-4f6c-475e-9d56-c69c67b207a8-kube-api-access-dnhg8\") pod \"machine-config-server-p7b2r\" (UID: \"800349eb-4f6c-475e-9d56-c69c67b207a8\") " pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.200831 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.201602 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.701588504 +0000 UTC m=+245.044060892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.204654 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.210763 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:23 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:23 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:23 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.210899 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.211386 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.212720 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tlh8\" (UniqueName: \"kubernetes.io/projected/1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40-kube-api-access-4tlh8\") pod \"catalog-operator-68c6474976-9hnv9\" (UID: \"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.239534 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.246489 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5p7g\" (UniqueName: \"kubernetes.io/projected/64f8c884-f5bf-4f9b-9c51-9e08d62100f3-kube-api-access-w5p7g\") pod \"olm-operator-6b444d44fb-75zml\" (UID: \"64f8c884-f5bf-4f9b-9c51-9e08d62100f3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.251052 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.253402 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckqqr\" (UniqueName: \"kubernetes.io/projected/c08aaa5f-186f-4f67-9332-d1b218a24943-kube-api-access-ckqqr\") pod \"service-ca-operator-777779d784-prpzs\" (UID: \"c08aaa5f-186f-4f67-9332-d1b218a24943\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.256261 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.264236 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.269497 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbbbz\" (UniqueName: \"kubernetes.io/projected/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-kube-api-access-cbbbz\") pod \"marketplace-operator-79b997595-c9j6s\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.280071 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4w7j2" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.282129 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3174c30-7f17-4ed8-b319-2ffc759ad14d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bv92v\" (UID: \"e3174c30-7f17-4ed8-b319-2ffc759ad14d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.294171 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-p7b2r" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.301968 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf2ml\" (UniqueName: \"kubernetes.io/projected/dac5b814-763c-42f7-a499-7a6afe1787b9-kube-api-access-qf2ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5d9d\" (UID: \"dac5b814-763c-42f7-a499-7a6afe1787b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.302685 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.303329 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.803309915 +0000 UTC m=+245.145782313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.330903 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5ngq\" (UniqueName: \"kubernetes.io/projected/93169b48-d875-458c-b7bd-ee2f4fd250f9-kube-api-access-z5ngq\") pod \"machine-config-controller-84d6567774-ksdsq\" (UID: \"93169b48-d875-458c-b7bd-ee2f4fd250f9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.335453 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5pntt"] Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.375568 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978fa61e-9c1b-4aae-8fef-e3763342994c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-44qdg\" (UID: \"978fa61e-9c1b-4aae-8fef-e3763342994c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.380182 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9d4l\" (UniqueName: \"kubernetes.io/projected/f821a876-eab1-4298-b06c-9560fb385085-kube-api-access-f9d4l\") pod \"oauth-openshift-558db77b4-jcrcb\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.400879 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqp4f\" (UniqueName: \"kubernetes.io/projected/49d76d7c-f89a-426a-8855-95b847669e5e-kube-api-access-cqp4f\") pod \"service-ca-9c57cc56f-2dgl2\" (UID: \"49d76d7c-f89a-426a-8855-95b847669e5e\") " pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.404508 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.405026 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvs8x\" (UniqueName: \"kubernetes.io/projected/93426711-c51f-4f98-b0aa-e0478447375d-kube-api-access-cvs8x\") pod \"kube-storage-version-migrator-operator-b67b599dd-8kkl8\" (UID: \"93426711-c51f-4f98-b0aa-e0478447375d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.405734 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.406320 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:23.90630647 +0000 UTC m=+245.248778868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.416304 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.426983 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfrz6\" (UniqueName: \"kubernetes.io/projected/22da2745-194a-4eea-bae4-923bdbaf6cab-kube-api-access-rfrz6\") pod \"dns-default-qjrwx\" (UID: \"22da2745-194a-4eea-bae4-923bdbaf6cab\") " pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.429783 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.436100 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.444687 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.450210 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.464145 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.482456 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.492051 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.498901 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509089 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509359 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509450 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509521 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509542 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509602 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509699 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509727 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509782 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.509806 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.511523 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-image-import-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.511676 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.01165889 +0000 UTC m=+245.354131288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.513515 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-config\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.514104 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/027f6e21-69f1-4111-85ec-f4f7486da3a5-images\") pod \"machine-api-operator-5694c8668f-p5c9r\" (UID: \"027f6e21-69f1-4111-85ec-f4f7486da3a5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.519157 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp"] Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.520710 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.521073 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.521613 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cfb68d8f-61a3-40d7-8e89-44084f690dc2-encryption-config\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.554277 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-audit\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.558783 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config\") pod \"controller-manager-879f6c89f-9xsqt\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.560299 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cfb68d8f-61a3-40d7-8e89-44084f690dc2-etcd-serving-ca\") pod \"apiserver-76f77b778f-pctzl\" (UID: \"cfb68d8f-61a3-40d7-8e89-44084f690dc2\") " pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.572285 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:23 crc kubenswrapper[4750]: W0309 18:29:23.609582 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod728fe455_8ea2_4cc2_afd0_981dae4d04b0.slice/crio-ea242f4aa0f93aff38eed8a68f85e2f5f1b56ad3dd19cfeaf78c0580ac087f08 WatchSource:0}: Error finding container ea242f4aa0f93aff38eed8a68f85e2f5f1b56ad3dd19cfeaf78c0580ac087f08: Status 404 returned error can't find the container with id ea242f4aa0f93aff38eed8a68f85e2f5f1b56ad3dd19cfeaf78c0580ac087f08 Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.611992 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.612407 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.112393224 +0000 UTC m=+245.454865622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.663955 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.684451 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.727270 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.728292 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.228268041 +0000 UTC m=+245.570740439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.734016 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.829728 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.830836 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.330820695 +0000 UTC m=+245.673293093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.849737 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" event={"ID":"90b2ac21-294a-40b6-b91e-8b189e5cca78","Type":"ContainerStarted","Data":"e6d033f993ffbfc6778036e0a427bd929de9dda6118e56a3b3d136bb0f606c08"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.862229 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j"] Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.868873 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" event={"ID":"f131dc29-3928-4ef8-a3e6-d59942a59ffd","Type":"ContainerStarted","Data":"c00ecb25baaa9e38e9b9f5bdc2ea4443d0b65074895177c4dc87b8c2d92889b9"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.868935 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" event={"ID":"f131dc29-3928-4ef8-a3e6-d59942a59ffd","Type":"ContainerStarted","Data":"463e304e1607e422055db4ca2498af4b9f38cc4823258f9ee7ca5288d4c56826"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.870984 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758"] Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.902192 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" event={"ID":"8d09c46d-08c7-436b-9427-49b3c9b5f875","Type":"ContainerStarted","Data":"f0cd3fee2a1e4e7a406f020a84f818e5be2360e553a80ce3a411e6744d66adb5"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.902255 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" event={"ID":"8d09c46d-08c7-436b-9427-49b3c9b5f875","Type":"ContainerStarted","Data":"41d69d181d557500705bb630b4d598d54e5085730706c9276b53cd96d9093ed7"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.908765 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-scg4c" event={"ID":"43705153-a050-4d95-ba46-792661ace7ae","Type":"ContainerStarted","Data":"a6addd6ff5536108a41349db82dadac701b190cf46b96abea34b69f2d09b4cc4"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.908856 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-scg4c" event={"ID":"43705153-a050-4d95-ba46-792661ace7ae","Type":"ContainerStarted","Data":"0e5d4fc5a680e9b3ef23ec7b73f752ee0f8cae623151afbefd98d0f886f8ae00"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.909337 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-scg4c" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.913484 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" event={"ID":"bb7ec7d3-ebab-4f3d-b14d-159046f5b354","Type":"ContainerStarted","Data":"2b2dfed6ea280784038c5a1bcba99fdcb39676cf45382cdedaa500d4a740d66d"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.913533 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" event={"ID":"bb7ec7d3-ebab-4f3d-b14d-159046f5b354","Type":"ContainerStarted","Data":"49d3396f0fbc795451ea27a718eb0549c10843c5d4717ff11d63cb498d9157ba"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.919937 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-scg4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.919997 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-scg4c" podUID="43705153-a050-4d95-ba46-792661ace7ae" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.935263 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:23 crc kubenswrapper[4750]: E0309 18:29:23.935991 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.435954588 +0000 UTC m=+245.778426996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.937260 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" event={"ID":"078d41c4-0ab6-4c47-9179-c332c66882c5","Type":"ContainerStarted","Data":"1446688359adbcc3140698511f4f653091c9cce450a9dbf39b7c9c33367728be"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.964784 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" event={"ID":"728fe455-8ea2-4cc2-afd0-981dae4d04b0","Type":"ContainerStarted","Data":"ea242f4aa0f93aff38eed8a68f85e2f5f1b56ad3dd19cfeaf78c0580ac087f08"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.976293 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" event={"ID":"3484441e-7900-421d-971a-14581e0d31ff","Type":"ContainerStarted","Data":"4c847f229a6fdd4cf465b37c1b4e37629032f266e31f462b63325f57a11db018"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.976392 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" event={"ID":"3484441e-7900-421d-971a-14581e0d31ff","Type":"ContainerStarted","Data":"0bc1a5e2195abeed5a081561b4f66295b772550080974e6db46c976f9379f0f2"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.989755 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" event={"ID":"2beaf3c7-025c-4bf7-b540-e0701c881860","Type":"ContainerStarted","Data":"54de8a9de6707533a5137e3687ec2c0a8770e04052dcad31dfb6d5b291bc040c"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.991117 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-p7b2r" event={"ID":"800349eb-4f6c-475e-9d56-c69c67b207a8","Type":"ContainerStarted","Data":"40f67d0adbe73e5b8681ae3505dcb4d259c09dbc1edda8428b6c320e7594cb72"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.993143 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" event={"ID":"b8865d33-f955-449d-bba5-9c42abf0e7e0","Type":"ContainerStarted","Data":"48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060"} Mar 09 18:29:23 crc kubenswrapper[4750]: I0309 18:29:23.994255 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.001645 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" event={"ID":"bd3871c8-50bf-434d-abb9-a457e15fba76","Type":"ContainerStarted","Data":"eb2beec25e89bee7bdb80619660ecfc98fa97a340e6a5fc14167215a9902e8d1"} Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.039729 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.047153 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.547127367 +0000 UTC m=+245.889599945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.141322 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.141654 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.641634931 +0000 UTC m=+245.984107329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.201421 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:24 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:24 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:24 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.201938 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:24 crc kubenswrapper[4750]: W0309 18:29:24.220840 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod483c734b_8a73_4053_b68c_256672b2f5c8.slice/crio-46e6310b931cf55f50340e4c0c05d35f78d9ee5de8847aa542c9ccd2a4590055 WatchSource:0}: Error finding container 46e6310b931cf55f50340e4c0c05d35f78d9ee5de8847aa542c9ccd2a4590055: Status 404 returned error can't find the container with id 46e6310b931cf55f50340e4c0c05d35f78d9ee5de8847aa542c9ccd2a4590055 Mar 09 18:29:24 crc kubenswrapper[4750]: W0309 18:29:24.224003 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6708629c_b5b3_4817_add5_f39c09860086.slice/crio-af1a13ff4b7b15302f54fab87c68e00d4d2d69b73673abd1c82cd341d365d6c6 WatchSource:0}: Error finding container af1a13ff4b7b15302f54fab87c68e00d4d2d69b73673abd1c82cd341d365d6c6: Status 404 returned error can't find the container with id af1a13ff4b7b15302f54fab87c68e00d4d2d69b73673abd1c82cd341d365d6c6 Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.244755 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.245122 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.745102079 +0000 UTC m=+246.087574477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.357489 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.357811 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.358650 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.858630302 +0000 UTC m=+246.201102690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.408261 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" podStartSLOduration=176.408233607 podStartE2EDuration="2m56.408233607s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:24.400097495 +0000 UTC m=+245.742569893" watchObservedRunningTime="2026-03-09 18:29:24.408233607 +0000 UTC m=+245.750706005" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.438912 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40328: no serving certificate available for the kubelet" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.462039 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.462392 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:24.962375028 +0000 UTC m=+246.304847426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.555353 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40342: no serving certificate available for the kubelet" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.569907 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.570341 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.070323678 +0000 UTC m=+246.412796076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.584375 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-54t4s" podStartSLOduration=176.584357232 podStartE2EDuration="2m56.584357232s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:24.541751757 +0000 UTC m=+245.884224155" watchObservedRunningTime="2026-03-09 18:29:24.584357232 +0000 UTC m=+245.926829630" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.671860 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.672680 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.172647385 +0000 UTC m=+246.515119783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.674713 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40358: no serving certificate available for the kubelet" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.681218 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-jn8bm" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.683850 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djq6g" podStartSLOduration=176.683786809 podStartE2EDuration="2m56.683786809s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:24.674829594 +0000 UTC m=+246.017301992" watchObservedRunningTime="2026-03-09 18:29:24.683786809 +0000 UTC m=+246.026259207" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.719679 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-97lvh" podStartSLOduration=176.719638679 podStartE2EDuration="2m56.719638679s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:24.718260692 +0000 UTC m=+246.060733090" watchObservedRunningTime="2026-03-09 18:29:24.719638679 +0000 UTC m=+246.062111077" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.730190 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-j5scj"] Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.760406 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" podStartSLOduration=175.760384393 podStartE2EDuration="2m55.760384393s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:24.759348875 +0000 UTC m=+246.101821283" watchObservedRunningTime="2026-03-09 18:29:24.760384393 +0000 UTC m=+246.102856801" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.764807 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8"] Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.773649 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.774107 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.274087687 +0000 UTC m=+246.616560085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.790422 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr"] Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.813181 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40368: no serving certificate available for the kubelet" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.851485 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc"] Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.876116 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9xsqt"] Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.880359 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.881298 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.381281818 +0000 UTC m=+246.723754216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.904936 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-bglg2" podStartSLOduration=176.904911044 podStartE2EDuration="2m56.904911044s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:24.883022585 +0000 UTC m=+246.225494983" watchObservedRunningTime="2026-03-09 18:29:24.904911044 +0000 UTC m=+246.247383442" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.949922 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40376: no serving certificate available for the kubelet" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.958343 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-scg4c" podStartSLOduration=176.958322254 podStartE2EDuration="2m56.958322254s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:24.95747073 +0000 UTC m=+246.299943128" watchObservedRunningTime="2026-03-09 18:29:24.958322254 +0000 UTC m=+246.300794652" Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.986701 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.986874 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.486842293 +0000 UTC m=+246.829314691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:24 crc kubenswrapper[4750]: I0309 18:29:24.987086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:24 crc kubenswrapper[4750]: E0309 18:29:24.987985 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.487969784 +0000 UTC m=+246.830442182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.036958 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk"] Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.066014 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp"] Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.078376 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" event={"ID":"6708629c-b5b3-4817-add5-f39c09860086","Type":"ContainerStarted","Data":"af1a13ff4b7b15302f54fab87c68e00d4d2d69b73673abd1c82cd341d365d6c6"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.078472 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40382: no serving certificate available for the kubelet" Mar 09 18:29:25 crc kubenswrapper[4750]: W0309 18:29:25.082431 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0ecde8f_d9fa_45e5_a508_981772436b4f.slice/crio-5f2cd1804ad59b35c14c42a91bba3c8b769ee5bb20bb1af6354d213391070e21 WatchSource:0}: Error finding container 5f2cd1804ad59b35c14c42a91bba3c8b769ee5bb20bb1af6354d213391070e21: Status 404 returned error can't find the container with id 5f2cd1804ad59b35c14c42a91bba3c8b769ee5bb20bb1af6354d213391070e21 Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.091778 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.092544 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.592518592 +0000 UTC m=+246.934991000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.094111 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" event={"ID":"332ab927-566e-4de9-bb3b-9784f123146b","Type":"ContainerStarted","Data":"1cbd379f0dfa3fc9b91b50fdcdc98e8f2779f03614423d1fca8bf44d78cff7ce"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.149313 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" event={"ID":"bd3871c8-50bf-434d-abb9-a457e15fba76","Type":"ContainerStarted","Data":"b3f71b9b9309d4ac59077adea84c56bdfcfa106ee9045c4103f350fe3540606a"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.154124 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40396: no serving certificate available for the kubelet" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.174221 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-p7b2r" event={"ID":"800349eb-4f6c-475e-9d56-c69c67b207a8","Type":"ContainerStarted","Data":"a0dff4ea363565f124b76e2b4a01f77bdc48a5597609a7e041d0af644eaede31"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.177823 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5w4wg" podStartSLOduration=177.177808323 podStartE2EDuration="2m57.177808323s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.159294006 +0000 UTC m=+246.501766404" watchObservedRunningTime="2026-03-09 18:29:25.177808323 +0000 UTC m=+246.520280721" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.184040 4750 generic.go:334] "Generic (PLEG): container finished" podID="f131dc29-3928-4ef8-a3e6-d59942a59ffd" containerID="c00ecb25baaa9e38e9b9f5bdc2ea4443d0b65074895177c4dc87b8c2d92889b9" exitCode=0 Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.184958 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" event={"ID":"f131dc29-3928-4ef8-a3e6-d59942a59ffd","Type":"ContainerDied","Data":"c00ecb25baaa9e38e9b9f5bdc2ea4443d0b65074895177c4dc87b8c2d92889b9"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.186544 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" event={"ID":"0e15e4fc-3419-4de4-b5db-6b10b1d4b962","Type":"ContainerStarted","Data":"d097165f221ab4c9f85c5886ae9c88d8d92d13b26dc8f6c38bb5eef13c5b012e"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.190373 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" event={"ID":"728fe455-8ea2-4cc2-afd0-981dae4d04b0","Type":"ContainerStarted","Data":"1a60133de179876d9b2771f0538acada7a6c41c89fef6c7735b6ba48f435b613"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.193537 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.195928 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.695912158 +0000 UTC m=+247.038384556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.197536 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" event={"ID":"483c734b-8a73-4053-b68c-256672b2f5c8","Type":"ContainerStarted","Data":"46e6310b931cf55f50340e4c0c05d35f78d9ee5de8847aa542c9ccd2a4590055"} Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.206932 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-scg4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.207013 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-scg4c" podUID="43705153-a050-4d95-ba46-792661ace7ae" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.296826 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.297276 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.797246578 +0000 UTC m=+247.139718986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.297737 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.300316 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:25 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:25 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:25 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.300385 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.301385 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.801365531 +0000 UTC m=+247.143837929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.321930 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40408: no serving certificate available for the kubelet" Mar 09 18:29:25 crc kubenswrapper[4750]: W0309 18:29:25.322107 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod653a7146_32eb_46ad_88bf_d16dddb2192f.slice/crio-c1f089fa564476dd63fc09b89c627875cdfca4d8d5495759f79c9af3dfb1a6a6 WatchSource:0}: Error finding container c1f089fa564476dd63fc09b89c627875cdfca4d8d5495759f79c9af3dfb1a6a6: Status 404 returned error can't find the container with id c1f089fa564476dd63fc09b89c627875cdfca4d8d5495759f79c9af3dfb1a6a6 Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.368035 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pfpff" podStartSLOduration=177.367999842 podStartE2EDuration="2m57.367999842s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.363079567 +0000 UTC m=+246.705551965" watchObservedRunningTime="2026-03-09 18:29:25.367999842 +0000 UTC m=+246.710472240" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.403012 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.403571 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:25.903549404 +0000 UTC m=+247.246021802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.438124 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fdg95" podStartSLOduration=177.438099348 podStartE2EDuration="2m57.438099348s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.434007886 +0000 UTC m=+246.776480284" watchObservedRunningTime="2026-03-09 18:29:25.438099348 +0000 UTC m=+246.780571746" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.478720 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-lll4q" podStartSLOduration=177.478697948 podStartE2EDuration="2m57.478697948s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.47807323 +0000 UTC m=+246.820545628" watchObservedRunningTime="2026-03-09 18:29:25.478697948 +0000 UTC m=+246.821170346" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.505281 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.505785 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.005746747 +0000 UTC m=+247.348219205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.516904 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cwxb8" podStartSLOduration=177.516876181 podStartE2EDuration="2m57.516876181s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.513453767 +0000 UTC m=+246.855926175" watchObservedRunningTime="2026-03-09 18:29:25.516876181 +0000 UTC m=+246.859348579" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.610531 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.611568 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.111529658 +0000 UTC m=+247.454002056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.712677 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.713054 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.213034353 +0000 UTC m=+247.555506751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.832326 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.833107 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.333074525 +0000 UTC m=+247.675546943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.854905 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-p7b2r" podStartSLOduration=6.85487916 podStartE2EDuration="6.85487916s" podCreationTimestamp="2026-03-09 18:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.854062878 +0000 UTC m=+247.196535276" watchObservedRunningTime="2026-03-09 18:29:25.85487916 +0000 UTC m=+247.197351558" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.859153 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" podStartSLOduration=177.859140266 podStartE2EDuration="2m57.859140266s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.81753309 +0000 UTC m=+247.160005488" watchObservedRunningTime="2026-03-09 18:29:25.859140266 +0000 UTC m=+247.201612664" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.907142 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-8wcqm" podStartSLOduration=177.907113918 podStartE2EDuration="2m57.907113918s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:25.904924468 +0000 UTC m=+247.247396866" watchObservedRunningTime="2026-03-09 18:29:25.907113918 +0000 UTC m=+247.249586316" Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.935985 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:25 crc kubenswrapper[4750]: E0309 18:29:25.936410 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.436394649 +0000 UTC m=+247.778867047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:25 crc kubenswrapper[4750]: I0309 18:29:25.967475 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.028706 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40414: no serving certificate available for the kubelet" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.034092 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.037169 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.037501 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.537483542 +0000 UTC m=+247.879955940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.051659 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.120777 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4w7j2"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.125947 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551348-g9vjb"] Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.138586 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.638568465 +0000 UTC m=+247.981040863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.138617 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.203811 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:26 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:26 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:26 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.204255 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.222255 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.252723 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" event={"ID":"f131dc29-3928-4ef8-a3e6-d59942a59ffd","Type":"ContainerStarted","Data":"4f48f00c463f2307d172639e87cc83368bb6ea1ad3049f1d309106fe7cfc659d"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.254871 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.255392 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.255625 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.755606923 +0000 UTC m=+248.098079321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.255791 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.256204 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.75619262 +0000 UTC m=+248.098665018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.257721 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" event={"ID":"d0ecde8f-d9fa-45e5-a508-981772436b4f","Type":"ContainerStarted","Data":"1aa681e696281cc67b93af1e699c7c67d083d7656749096e07580423b92d0786"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.257756 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" event={"ID":"d0ecde8f-d9fa-45e5-a508-981772436b4f","Type":"ContainerStarted","Data":"5f2cd1804ad59b35c14c42a91bba3c8b769ee5bb20bb1af6354d213391070e21"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.279037 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c9j6s"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.280882 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.283144 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jcrcb"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.285836 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9xsqt"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.296153 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" podStartSLOduration=178.296128131 podStartE2EDuration="2m58.296128131s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:26.293720436 +0000 UTC m=+247.636192834" watchObservedRunningTime="2026-03-09 18:29:26.296128131 +0000 UTC m=+247.638600529" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.298082 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" event={"ID":"653a7146-32eb-46ad-88bf-d16dddb2192f","Type":"ContainerStarted","Data":"e59fe0f70bcdda9a3a137c7b15c7bee39584acb4f9a054c83d15fe620185449e"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.298130 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" event={"ID":"653a7146-32eb-46ad-88bf-d16dddb2192f","Type":"ContainerStarted","Data":"c1f089fa564476dd63fc09b89c627875cdfca4d8d5495759f79c9af3dfb1a6a6"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.298496 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.303336 4750 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-hlcdk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.303430 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" podUID="653a7146-32eb-46ad-88bf-d16dddb2192f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.320521 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" event={"ID":"0e15e4fc-3419-4de4-b5db-6b10b1d4b962","Type":"ContainerStarted","Data":"ab9b71add30a3b66476af5e1f63728d9ed5be4bcfcb26da8633cb42b172be784"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.321389 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.323564 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" podStartSLOduration=178.32352302 podStartE2EDuration="2m58.32352302s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:26.313186118 +0000 UTC m=+247.655658516" watchObservedRunningTime="2026-03-09 18:29:26.32352302 +0000 UTC m=+247.665995418" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.333338 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" event={"ID":"a08fcb79-42d8-47c2-940e-92c4ff4f2f61","Type":"ContainerStarted","Data":"a447fc03882759db56727d5053167aa55d24d12a915c5d97fc7a09caddb70110"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.343973 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" event={"ID":"36c2456b-5e37-4a71-9c55-7e1bd2182512","Type":"ContainerStarted","Data":"ae61f95bba999a48195e22c4a251ba3d680882b4f11bc4a870da30215dee9f51"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.344476 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" event={"ID":"36c2456b-5e37-4a71-9c55-7e1bd2182512","Type":"ContainerStarted","Data":"5c5043cd759aedc16e6845361c363f59cbd5c23e45fa727f24c66790627fd4b9"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.354747 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" podStartSLOduration=177.354730183 podStartE2EDuration="2m57.354730183s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:26.344189905 +0000 UTC m=+247.686662303" watchObservedRunningTime="2026-03-09 18:29:26.354730183 +0000 UTC m=+247.697202581" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.357404 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" event={"ID":"6708629c-b5b3-4817-add5-f39c09860086","Type":"ContainerStarted","Data":"f85b5c377d4a6c50b14c93ed4c3b42d226ae7b2794ec9dc87a52883777a2c933"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.358504 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.359047 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq"] Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.385089 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.885066642 +0000 UTC m=+248.227539040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.385168 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-prpzs"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.388547 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" podStartSLOduration=177.388513216 podStartE2EDuration="2m57.388513216s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:26.376024385 +0000 UTC m=+247.718496783" watchObservedRunningTime="2026-03-09 18:29:26.388513216 +0000 UTC m=+247.730985614" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.389441 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4w7j2" event={"ID":"71bbf2b6-cac6-4959-9124-607c988da153","Type":"ContainerStarted","Data":"52e7569b10f6720f1e06613559b2f4ec826ea1e02858dd9b351f487c1c826c40"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.400485 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" event={"ID":"64f8c884-f5bf-4f9b-9c51-9e08d62100f3","Type":"ContainerStarted","Data":"41152260b5e6fa24201293691f78fff1445c02dfafaa7d451cfd98606cdd680a"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.405717 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.414413 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.428017 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5c9r"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.428104 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" event={"ID":"332ab927-566e-4de9-bb3b-9784f123146b","Type":"ContainerStarted","Data":"d3555d02998f50b1ef9e9e8c71f145d94ce841acb36d411adc2c18dc7c5d2b4c"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.474214 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.474447 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pctzl"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.475580 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" podStartSLOduration=178.475550406 podStartE2EDuration="2m58.475550406s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:26.414049535 +0000 UTC m=+247.756521933" watchObservedRunningTime="2026-03-09 18:29:26.475550406 +0000 UTC m=+247.818022804" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.484932 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.486358 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:26.98633886 +0000 UTC m=+248.328811328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.486413 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2dgl2"] Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.488316 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qg758" podStartSLOduration=178.488291844 podStartE2EDuration="2m58.488291844s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:26.43877518 +0000 UTC m=+247.781247598" watchObservedRunningTime="2026-03-09 18:29:26.488291844 +0000 UTC m=+247.830764242" Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.492722 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qjrwx"] Mar 09 18:29:26 crc kubenswrapper[4750]: W0309 18:29:26.522269 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc08aaa5f_186f_4f67_9332_d1b218a24943.slice/crio-8488c6209ddb6d6079bd2f2e57dc4734cc3e7b386b89000f675ad0b4bd467ec2 WatchSource:0}: Error finding container 8488c6209ddb6d6079bd2f2e57dc4734cc3e7b386b89000f675ad0b4bd467ec2: Status 404 returned error can't find the container with id 8488c6209ddb6d6079bd2f2e57dc4734cc3e7b386b89000f675ad0b4bd467ec2 Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.526854 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" event={"ID":"90b2ac21-294a-40b6-b91e-8b189e5cca78","Type":"ContainerStarted","Data":"ea41da53d676d565624b42dfb2f9010b49abac2651fde376c72741a705078bd3"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.544925 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cj8qp" event={"ID":"728fe455-8ea2-4cc2-afd0-981dae4d04b0","Type":"ContainerStarted","Data":"d8be22d835140eb9b8a86e6e4c59f7402943cd771c14503fccdccb1e4cdf6999"} Mar 09 18:29:26 crc kubenswrapper[4750]: W0309 18:29:26.559023 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod978fa61e_9c1b_4aae_8fef_e3763342994c.slice/crio-eba35f7ec1ac2d5b7039b8b6c6d1146df1f51ee9063b4207aa05de2049fa8c3d WatchSource:0}: Error finding container eba35f7ec1ac2d5b7039b8b6c6d1146df1f51ee9063b4207aa05de2049fa8c3d: Status 404 returned error can't find the container with id eba35f7ec1ac2d5b7039b8b6c6d1146df1f51ee9063b4207aa05de2049fa8c3d Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.564140 4750 generic.go:334] "Generic (PLEG): container finished" podID="483c734b-8a73-4053-b68c-256672b2f5c8" containerID="a3076b9f092b582b0df5c853ec62d4eb89b57e38a5e79f2ad3868e11aea5c54c" exitCode=0 Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.564264 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" event={"ID":"483c734b-8a73-4053-b68c-256672b2f5c8","Type":"ContainerDied","Data":"a3076b9f092b582b0df5c853ec62d4eb89b57e38a5e79f2ad3868e11aea5c54c"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.573269 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" event={"ID":"dac5b814-763c-42f7-a499-7a6afe1787b9","Type":"ContainerStarted","Data":"83fc1beaa4dcafda3e3cc1c36fc9c46c23862a332061e3cbde745036fb951ac7"} Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.587679 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.588328 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.088304978 +0000 UTC m=+248.430777376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: W0309 18:29:26.605815 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod027f6e21_69f1_4111_85ec_f4f7486da3a5.slice/crio-1eaf327f680a6df86cc50cb68323b85c4478a976eeb7ef378dd4b5b80c247d46 WatchSource:0}: Error finding container 1eaf327f680a6df86cc50cb68323b85c4478a976eeb7ef378dd4b5b80c247d46: Status 404 returned error can't find the container with id 1eaf327f680a6df86cc50cb68323b85c4478a976eeb7ef378dd4b5b80c247d46 Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.607071 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" podUID="b8865d33-f955-449d-bba5-9c42abf0e7e0" containerName="route-controller-manager" containerID="cri-o://48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060" gracePeriod=30 Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.607199 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" event={"ID":"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3","Type":"ContainerStarted","Data":"118c7b4e2cb14414d90ce5410e1efa45c4867e5089868d29ab7dbba8a17f8928"} Mar 09 18:29:26 crc kubenswrapper[4750]: W0309 18:29:26.661914 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49d76d7c_f89a_426a_8855_95b847669e5e.slice/crio-39503203e0e5d1181d13be7e17f014b54455122c7177557a303d77dca5663453 WatchSource:0}: Error finding container 39503203e0e5d1181d13be7e17f014b54455122c7177557a303d77dca5663453: Status 404 returned error can't find the container with id 39503203e0e5d1181d13be7e17f014b54455122c7177557a303d77dca5663453 Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.691887 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.694822 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.194806649 +0000 UTC m=+248.537279047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.793207 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.793825 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.293787784 +0000 UTC m=+248.636260182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.894755 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.895486 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.395459583 +0000 UTC m=+248.737932141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:26 crc kubenswrapper[4750]: I0309 18:29:26.996262 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:26 crc kubenswrapper[4750]: E0309 18:29:26.996635 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.496618309 +0000 UTC m=+248.839090707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.098788 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.099394 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.599365557 +0000 UTC m=+248.941837995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.200420 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.200832 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:27 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:27 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:27 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.200892 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.201045 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.701027006 +0000 UTC m=+249.043499404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.303864 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.304285 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.804267258 +0000 UTC m=+249.146739656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.348179 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.363197 4750 ???:1] "http: TLS handshake error from 192.168.126.11:40426: no serving certificate available for the kubelet" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.404929 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.405037 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-config\") pod \"b8865d33-f955-449d-bba5-9c42abf0e7e0\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.405072 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-client-ca\") pod \"b8865d33-f955-449d-bba5-9c42abf0e7e0\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.405102 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvwh9\" (UniqueName: \"kubernetes.io/projected/b8865d33-f955-449d-bba5-9c42abf0e7e0-kube-api-access-jvwh9\") pod \"b8865d33-f955-449d-bba5-9c42abf0e7e0\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.405228 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8865d33-f955-449d-bba5-9c42abf0e7e0-serving-cert\") pod \"b8865d33-f955-449d-bba5-9c42abf0e7e0\" (UID: \"b8865d33-f955-449d-bba5-9c42abf0e7e0\") " Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.407244 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:27.907220423 +0000 UTC m=+249.249692821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.407764 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-client-ca" (OuterVolumeSpecName: "client-ca") pod "b8865d33-f955-449d-bba5-9c42abf0e7e0" (UID: "b8865d33-f955-449d-bba5-9c42abf0e7e0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.407952 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-config" (OuterVolumeSpecName: "config") pod "b8865d33-f955-449d-bba5-9c42abf0e7e0" (UID: "b8865d33-f955-449d-bba5-9c42abf0e7e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.437193 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79"] Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.437585 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8865d33-f955-449d-bba5-9c42abf0e7e0" containerName="route-controller-manager" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.437615 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8865d33-f955-449d-bba5-9c42abf0e7e0" containerName="route-controller-manager" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.437749 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8865d33-f955-449d-bba5-9c42abf0e7e0" containerName="route-controller-manager" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.438172 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79"] Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.438414 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.441895 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8865d33-f955-449d-bba5-9c42abf0e7e0-kube-api-access-jvwh9" (OuterVolumeSpecName: "kube-api-access-jvwh9") pod "b8865d33-f955-449d-bba5-9c42abf0e7e0" (UID: "b8865d33-f955-449d-bba5-9c42abf0e7e0"). InnerVolumeSpecName "kube-api-access-jvwh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.444263 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8865d33-f955-449d-bba5-9c42abf0e7e0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b8865d33-f955-449d-bba5-9c42abf0e7e0" (UID: "b8865d33-f955-449d-bba5-9c42abf0e7e0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.510514 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/178203e6-eb24-4f4f-b794-b11b496778e4-serving-cert\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.510623 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.511125 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-config\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.511450 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.01143246 +0000 UTC m=+249.353904858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.511694 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-client-ca\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.511758 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrt2s\" (UniqueName: \"kubernetes.io/projected/178203e6-eb24-4f4f-b794-b11b496778e4-kube-api-access-jrt2s\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.511888 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8865d33-f955-449d-bba5-9c42abf0e7e0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.511915 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.511931 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8865d33-f955-449d-bba5-9c42abf0e7e0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.511945 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvwh9\" (UniqueName: \"kubernetes.io/projected/b8865d33-f955-449d-bba5-9c42abf0e7e0-kube-api-access-jvwh9\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.613364 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.613718 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrt2s\" (UniqueName: \"kubernetes.io/projected/178203e6-eb24-4f4f-b794-b11b496778e4-kube-api-access-jrt2s\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.613868 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.11382635 +0000 UTC m=+249.456298748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.614086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/178203e6-eb24-4f4f-b794-b11b496778e4-serving-cert\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.614201 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.614228 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-config\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.614247 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-client-ca\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.615170 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-client-ca\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.616185 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.116169934 +0000 UTC m=+249.458642332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.617205 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-config\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.624096 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/178203e6-eb24-4f4f-b794-b11b496778e4-serving-cert\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.688829 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrt2s\" (UniqueName: \"kubernetes.io/projected/178203e6-eb24-4f4f-b794-b11b496778e4-kube-api-access-jrt2s\") pod \"route-controller-manager-66dc67c9bf-vzb79\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.694936 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4w7j2" event={"ID":"71bbf2b6-cac6-4959-9124-607c988da153","Type":"ContainerStarted","Data":"044810eb05d0e93eb115a9a47e1772ef8eda771a51a10f964d563fb3e95fe3ed"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.698238 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" event={"ID":"49d76d7c-f89a-426a-8855-95b847669e5e","Type":"ContainerStarted","Data":"99f37c844ba2cb431a8f52a481ddd867975be898f4d4007829a84ab61ab4a135"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.698377 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" event={"ID":"49d76d7c-f89a-426a-8855-95b847669e5e","Type":"ContainerStarted","Data":"39503203e0e5d1181d13be7e17f014b54455122c7177557a303d77dca5663453"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.710282 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" event={"ID":"e3174c30-7f17-4ed8-b319-2ffc759ad14d","Type":"ContainerStarted","Data":"6734275e617bb414c69944100b36221cc013cb6c67686e8b3c119d8e60a8229e"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.714221 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" event={"ID":"f821a876-eab1-4298-b06c-9560fb385085","Type":"ContainerStarted","Data":"12dca8646a722a54a0300fa6928c9ce622ac8d87b754af1c3c6c3a386ff967a5"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.715490 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.716508 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.216488916 +0000 UTC m=+249.558961314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.723682 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" event={"ID":"93426711-c51f-4f98-b0aa-e0478447375d","Type":"ContainerStarted","Data":"cd9e4038ff32a4d2a39d0df83c829a171dcf14b599466a4bcbc6a81389b9ae5b"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.723723 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" event={"ID":"93426711-c51f-4f98-b0aa-e0478447375d","Type":"ContainerStarted","Data":"7e31f937a3e62802e1a410ed9ff8688e319fc2d9b0c9316864ebdb7ba482c2f0"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.743458 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" event={"ID":"332ab927-566e-4de9-bb3b-9784f123146b","Type":"ContainerStarted","Data":"dc4b041a2fdac95bd6989fb8f60b178076bc73e2eac9b4b08f8b060f3d542a44"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.750044 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-2dgl2" podStartSLOduration=178.750017183 podStartE2EDuration="2m58.750017183s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:27.748149211 +0000 UTC m=+249.090621609" watchObservedRunningTime="2026-03-09 18:29:27.750017183 +0000 UTC m=+249.092489571" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.751221 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4w7j2" podStartSLOduration=8.751215115 podStartE2EDuration="8.751215115s" podCreationTimestamp="2026-03-09 18:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:27.723125198 +0000 UTC m=+249.065597596" watchObservedRunningTime="2026-03-09 18:29:27.751215115 +0000 UTC m=+249.093687513" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.790865 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.797671 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-j5scj" podStartSLOduration=178.793628474 podStartE2EDuration="2m58.793628474s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:27.791712263 +0000 UTC m=+249.134184671" watchObservedRunningTime="2026-03-09 18:29:27.793628474 +0000 UTC m=+249.136100872" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.801071 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" event={"ID":"cfb68d8f-61a3-40d7-8e89-44084f690dc2","Type":"ContainerStarted","Data":"c46557fe68e7b52e15163440fbc7a08c5f3ab1597b413355e1b6a42ee122e6b1"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.827130 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.828684 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8kkl8" podStartSLOduration=179.828661962 podStartE2EDuration="2m59.828661962s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:27.826142503 +0000 UTC m=+249.168614901" watchObservedRunningTime="2026-03-09 18:29:27.828661962 +0000 UTC m=+249.171134350" Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.829575 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.329553077 +0000 UTC m=+249.672025465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.834438 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" event={"ID":"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40","Type":"ContainerStarted","Data":"c1e1c71ce13ffe261bfce3af3d2ec0cc903f17e07cbd025a358affe749b6e736"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.834510 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" event={"ID":"1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40","Type":"ContainerStarted","Data":"84829e615ce38b8f1e0505e963cccdd45ffd3bad048babb719283e067f85d58f"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.834867 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.858036 4750 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9hnv9 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.858115 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" podUID="1f3bdc2a-e6a6-4b85-8dc1-8f2fa76a6f40" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.879892 4750 generic.go:334] "Generic (PLEG): container finished" podID="b8865d33-f955-449d-bba5-9c42abf0e7e0" containerID="48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060" exitCode=0 Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.880092 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" event={"ID":"b8865d33-f955-449d-bba5-9c42abf0e7e0","Type":"ContainerDied","Data":"48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.880143 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" event={"ID":"b8865d33-f955-449d-bba5-9c42abf0e7e0","Type":"ContainerDied","Data":"f7f19f6f861ebbc1b0f4af95b0f0db615cade4a08c1b3cdb2804bc1001887e92"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.880166 4750 scope.go:117] "RemoveContainer" containerID="48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.880388 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.908374 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" event={"ID":"978fa61e-9c1b-4aae-8fef-e3763342994c","Type":"ContainerStarted","Data":"eba35f7ec1ac2d5b7039b8b6c6d1146df1f51ee9063b4207aa05de2049fa8c3d"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.911600 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" event={"ID":"39ea59b4-900a-43ce-9cf5-b5d029ff0ada","Type":"ContainerStarted","Data":"0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.911686 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" event={"ID":"39ea59b4-900a-43ce-9cf5-b5d029ff0ada","Type":"ContainerStarted","Data":"bacaa8f1a6265292ac23fa9afb204b8509f66bcf22d8e01c8e5a17678b7db826"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.913052 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.928463 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.930512 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.430479976 +0000 UTC m=+249.772952524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.947317 4750 scope.go:117] "RemoveContainer" containerID="48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.948068 4750 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-c9j6s container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.948124 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.950092 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" podStartSLOduration=178.950076711 podStartE2EDuration="2m58.950076711s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:27.8728143 +0000 UTC m=+249.215286718" watchObservedRunningTime="2026-03-09 18:29:27.950076711 +0000 UTC m=+249.292549099" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.955697 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" podStartSLOduration=178.955666774 podStartE2EDuration="2m58.955666774s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:27.944422857 +0000 UTC m=+249.286895255" watchObservedRunningTime="2026-03-09 18:29:27.955666774 +0000 UTC m=+249.298139172" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.957705 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" event={"ID":"027f6e21-69f1-4111-85ec-f4f7486da3a5","Type":"ContainerStarted","Data":"ac2c100b9fd9cce5cab1c1bc33017b45927122165dfaf612d97882e4a3384eb0"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.957754 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" event={"ID":"027f6e21-69f1-4111-85ec-f4f7486da3a5","Type":"ContainerStarted","Data":"1eaf327f680a6df86cc50cb68323b85c4478a976eeb7ef378dd4b5b80c247d46"} Mar 09 18:29:27 crc kubenswrapper[4750]: E0309 18:29:27.960649 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060\": container with ID starting with 48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060 not found: ID does not exist" containerID="48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.960717 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060"} err="failed to get container status \"48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060\": rpc error: code = NotFound desc = could not find container \"48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060\": container with ID starting with 48f51319c6ea42bb57951cb4b52c8f9ad9a66f8d70f08d09e5c5f8e21f03c060 not found: ID does not exist" Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.976166 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" event={"ID":"c08aaa5f-186f-4f67-9332-d1b218a24943","Type":"ContainerStarted","Data":"0275dc2b4c2c70c158011a1320342762d950bdae0a36351abf5ae0c6e15c4e7f"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.976233 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" event={"ID":"c08aaa5f-186f-4f67-9332-d1b218a24943","Type":"ContainerStarted","Data":"8488c6209ddb6d6079bd2f2e57dc4734cc3e7b386b89000f675ad0b4bd467ec2"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.977825 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc"] Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.984785 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" event={"ID":"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3","Type":"ContainerStarted","Data":"79c242bccfaf100f9ca56e05935eec7f8af3fa83e08f459aad056d43ecbe5390"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.984844 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" event={"ID":"57e3e8d3-dfc2-42db-b4ed-e790c73cb5f3","Type":"ContainerStarted","Data":"0cbbc5e09a17ff3f949424bd4af7d6fd693bb1949ef59d0acb93af4f21920fb9"} Mar 09 18:29:27 crc kubenswrapper[4750]: I0309 18:29:27.986755 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2sfc"] Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:27.997611 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b84rp" event={"ID":"36c2456b-5e37-4a71-9c55-7e1bd2182512","Type":"ContainerStarted","Data":"c463ab7193bdcba93ccc35a8c80b8ecf6283277340805746710962449eb502c9"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.019265 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" event={"ID":"dac5b814-763c-42f7-a499-7a6afe1787b9","Type":"ContainerStarted","Data":"db654690028c4d86d1a77d45af386db569b41635cb3dab4c0e8689a5873d290b"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.030428 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.033644 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.533622954 +0000 UTC m=+249.876095352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.046415 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prpzs" podStartSLOduration=179.046390653 podStartE2EDuration="2m59.046390653s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:28.022147421 +0000 UTC m=+249.364619829" watchObservedRunningTime="2026-03-09 18:29:28.046390653 +0000 UTC m=+249.388863051" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.047369 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5d9d" podStartSLOduration=180.04736415 podStartE2EDuration="3m0.04736415s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:28.040477092 +0000 UTC m=+249.382949490" watchObservedRunningTime="2026-03-09 18:29:28.04736415 +0000 UTC m=+249.389836548" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.055930 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" event={"ID":"64f8c884-f5bf-4f9b-9c51-9e08d62100f3","Type":"ContainerStarted","Data":"1846293fd9b80e872b47e667512e3ba668b3ea7a6ff7002d94bb5645e59eca9a"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.062205 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.089721 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-94ztz" podStartSLOduration=179.089691097 podStartE2EDuration="2m59.089691097s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:28.072933429 +0000 UTC m=+249.415405827" watchObservedRunningTime="2026-03-09 18:29:28.089691097 +0000 UTC m=+249.432163495" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.103920 4750 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-75zml container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" start-of-body= Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.103981 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" podUID="64f8c884-f5bf-4f9b-9c51-9e08d62100f3" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.112822 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" podStartSLOduration=179.112788918 podStartE2EDuration="2m59.112788918s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:28.103430672 +0000 UTC m=+249.445903070" watchObservedRunningTime="2026-03-09 18:29:28.112788918 +0000 UTC m=+249.455261316" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.122473 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" event={"ID":"93169b48-d875-458c-b7bd-ee2f4fd250f9","Type":"ContainerStarted","Data":"3012c75cf01860ca5d7a8d95f14c34ac6173fc4181066107e4e80ff5b0d5d12d"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.122564 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" event={"ID":"93169b48-d875-458c-b7bd-ee2f4fd250f9","Type":"ContainerStarted","Data":"88311fc5f390ed70a289fbdd7b6bb0ef29913bcbdc2eeece69c6a1aff9a1ef21"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.141379 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.147298 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.647274992 +0000 UTC m=+249.989747390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.158876 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" podStartSLOduration=179.158858008 podStartE2EDuration="2m59.158858008s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:28.158129248 +0000 UTC m=+249.500601646" watchObservedRunningTime="2026-03-09 18:29:28.158858008 +0000 UTC m=+249.501330406" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.163333 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qjrwx" event={"ID":"22da2745-194a-4eea-bae4-923bdbaf6cab","Type":"ContainerStarted","Data":"6a21c951df3cea0c818ee2115e254caf8fc7642e6fc4b154ce35794242ecf4e9"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.191849 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" event={"ID":"06164218-c618-4751-8f8c-d21b71f7bda5","Type":"ContainerStarted","Data":"94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.191937 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" event={"ID":"06164218-c618-4751-8f8c-d21b71f7bda5","Type":"ContainerStarted","Data":"93027063b8fae75ab25320a11e611e0f6f0dbf6d14169c918dc2c29723a74600"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.192123 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" podUID="06164218-c618-4751-8f8c-d21b71f7bda5" containerName="controller-manager" containerID="cri-o://94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4" gracePeriod=30 Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.192756 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.198908 4750 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9xsqt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.198972 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" podUID="06164218-c618-4751-8f8c-d21b71f7bda5" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.204778 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:28 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:28 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:28 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.204829 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.229955 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" podStartSLOduration=180.229931271 podStartE2EDuration="3m0.229931271s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:28.22809308 +0000 UTC m=+249.570565478" watchObservedRunningTime="2026-03-09 18:29:28.229931271 +0000 UTC m=+249.572403669" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.241001 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" event={"ID":"0e15e4fc-3419-4de4-b5db-6b10b1d4b962","Type":"ContainerStarted","Data":"2e5585542543be2a8c67ec98f271945ca83db39b7311f01c723de81546af0810"} Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.252093 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.253479 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.753463173 +0000 UTC m=+250.095935571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.356507 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.356901 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.85687303 +0000 UTC m=+250.199345428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.357184 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.360171 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.86015744 +0000 UTC m=+250.202629838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.401918 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hlcdk" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.412485 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79"] Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.459558 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.460870 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:28.960845752 +0000 UTC m=+250.303318150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.566273 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.566711 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.066694226 +0000 UTC m=+250.409166624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.667242 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.667662 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.167628195 +0000 UTC m=+250.510100593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.778094 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.778438 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.278424803 +0000 UTC m=+250.620897201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.879261 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-9xsqt_06164218-c618-4751-8f8c-d21b71f7bda5/controller-manager/0.log" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.879343 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.879498 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.880155 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.380139034 +0000 UTC m=+250.722611432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.980675 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config\") pod \"06164218-c618-4751-8f8c-d21b71f7bda5\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.980756 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles\") pod \"06164218-c618-4751-8f8c-d21b71f7bda5\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.981826 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "06164218-c618-4751-8f8c-d21b71f7bda5" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.981998 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config" (OuterVolumeSpecName: "config") pod "06164218-c618-4751-8f8c-d21b71f7bda5" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.982077 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2jtf\" (UniqueName: \"kubernetes.io/projected/06164218-c618-4751-8f8c-d21b71f7bda5-kube-api-access-v2jtf\") pod \"06164218-c618-4751-8f8c-d21b71f7bda5\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.982134 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert\") pod \"06164218-c618-4751-8f8c-d21b71f7bda5\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.982152 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca\") pod \"06164218-c618-4751-8f8c-d21b71f7bda5\" (UID: \"06164218-c618-4751-8f8c-d21b71f7bda5\") " Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.982907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca" (OuterVolumeSpecName: "client-ca") pod "06164218-c618-4751-8f8c-d21b71f7bda5" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.984059 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.984238 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.984254 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:28 crc kubenswrapper[4750]: I0309 18:29:28.984262 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06164218-c618-4751-8f8c-d21b71f7bda5-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:28 crc kubenswrapper[4750]: E0309 18:29:28.984529 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.484509777 +0000 UTC m=+250.826982245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.000278 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "06164218-c618-4751-8f8c-d21b71f7bda5" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.002973 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06164218-c618-4751-8f8c-d21b71f7bda5-kube-api-access-v2jtf" (OuterVolumeSpecName: "kube-api-access-v2jtf") pod "06164218-c618-4751-8f8c-d21b71f7bda5" (UID: "06164218-c618-4751-8f8c-d21b71f7bda5"). InnerVolumeSpecName "kube-api-access-v2jtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.085104 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.085370 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.585332023 +0000 UTC m=+250.927804421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.085429 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.085734 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2jtf\" (UniqueName: \"kubernetes.io/projected/06164218-c618-4751-8f8c-d21b71f7bda5-kube-api-access-v2jtf\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.085762 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06164218-c618-4751-8f8c-d21b71f7bda5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.086306 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.586102263 +0000 UTC m=+250.928574661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.187752 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.188327 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.688245705 +0000 UTC m=+251.030718103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.189263 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.189728 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.689716995 +0000 UTC m=+251.032189573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.201336 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:29 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:29 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:29 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.201522 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.291052 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.291691 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.291968 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.79194647 +0000 UTC m=+251.134418868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.296043 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.305671 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" event={"ID":"483c734b-8a73-4053-b68c-256672b2f5c8","Type":"ContainerStarted","Data":"4d6fe52fe61477223df13393564819ca7f8e934aa30c225039f4e6f5f28db9d0"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.305838 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.327509 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qjrwx" event={"ID":"22da2745-194a-4eea-bae4-923bdbaf6cab","Type":"ContainerStarted","Data":"9fd772ef3213021787e9bcc85ee2a4ccab7123ea2a400fd5c5da3eb92fe2c7ed"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.327559 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qjrwx" event={"ID":"22da2745-194a-4eea-bae4-923bdbaf6cab","Type":"ContainerStarted","Data":"4ec6f276d92d9dceb40ea40ce44313fb5d87bb54293493071d96d45c408fb17e"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.328849 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.331591 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" podStartSLOduration=180.331573123 podStartE2EDuration="3m0.331573123s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:29.330247166 +0000 UTC m=+250.672719564" watchObservedRunningTime="2026-03-09 18:29:29.331573123 +0000 UTC m=+250.674045521" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.354994 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-qjrwx" podStartSLOduration=10.354973053 podStartE2EDuration="10.354973053s" podCreationTimestamp="2026-03-09 18:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:29.353986145 +0000 UTC m=+250.696458543" watchObservedRunningTime="2026-03-09 18:29:29.354973053 +0000 UTC m=+250.697445451" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.390138 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8865d33-f955-449d-bba5-9c42abf0e7e0" path="/var/lib/kubelet/pods/b8865d33-f955-449d-bba5-9c42abf0e7e0/volumes" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.391189 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.391214 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" event={"ID":"178203e6-eb24-4f4f-b794-b11b496778e4","Type":"ContainerStarted","Data":"0ca0155e7ea9f0108cf9a825a6a8d9b2a6513ff48fa06e8ff5cde596fe50e039"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.391233 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" event={"ID":"178203e6-eb24-4f4f-b794-b11b496778e4","Type":"ContainerStarted","Data":"d697a49b3584135834944a4fad0af885c5515316a20be8c8fb691266dd57a860"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.402683 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.402813 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.402836 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.402853 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.402909 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.403463 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:29.903446368 +0000 UTC m=+251.245918766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.412978 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.413210 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.422269 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.423196 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" event={"ID":"027f6e21-69f1-4111-85ec-f4f7486da3a5","Type":"ContainerStarted","Data":"b4af4975a5afcaa867f433f5cfdab8edf5bec146747036fc71a6ed2b00ef4e1a"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.427248 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.432893 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.433768 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a32a985-7c9a-4737-86db-d46edf0d0565-metrics-certs\") pod \"network-metrics-daemon-wl7w6\" (UID: \"6a32a985-7c9a-4737-86db-d46edf0d0565\") " pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.448166 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.456889 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" event={"ID":"90b2ac21-294a-40b6-b91e-8b189e5cca78","Type":"ContainerStarted","Data":"b21ec8363c390a71916d61a95b6fcedbf0652b500756de3e5e820374297f3646"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.456990 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.506086 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" event={"ID":"e3174c30-7f17-4ed8-b319-2ffc759ad14d","Type":"ContainerStarted","Data":"8e4e6c3c009df84b3bef3f418d2976d743caff018eff6149f355613df51c4338"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.507193 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.508712 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.008690005 +0000 UTC m=+251.351162403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.530215 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" event={"ID":"978fa61e-9c1b-4aae-8fef-e3763342994c","Type":"ContainerStarted","Data":"2b9c49a231d95739d757de4f45889bc407a5a724342eb3fc0c699f74ba99f7d9"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.547697 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" event={"ID":"f821a876-eab1-4298-b06c-9560fb385085","Type":"ContainerStarted","Data":"26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.548976 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.554263 4750 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jcrcb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused" start-of-body= Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.554320 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" podUID="f821a876-eab1-4298-b06c-9560fb385085" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.565801 4750 generic.go:334] "Generic (PLEG): container finished" podID="cfb68d8f-61a3-40d7-8e89-44084f690dc2" containerID="0cb23e7859ad94f8fd26c7a37a7ba77e38152a478ea4f1787b9b3840154bb6ba" exitCode=0 Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.565991 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" event={"ID":"cfb68d8f-61a3-40d7-8e89-44084f690dc2","Type":"ContainerDied","Data":"0cb23e7859ad94f8fd26c7a37a7ba77e38152a478ea4f1787b9b3840154bb6ba"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.605365 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ksdsq" event={"ID":"93169b48-d875-458c-b7bd-ee2f4fd250f9","Type":"ContainerStarted","Data":"9b3be87d04c19f4253a0375d1e27c4df7b823948dd2f820e7baf042778c9320c"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.608238 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-9xsqt_06164218-c618-4751-8f8c-d21b71f7bda5/controller-manager/0.log" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.608275 4750 generic.go:334] "Generic (PLEG): container finished" podID="06164218-c618-4751-8f8c-d21b71f7bda5" containerID="94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4" exitCode=2 Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.609168 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.609553 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" event={"ID":"06164218-c618-4751-8f8c-d21b71f7bda5","Type":"ContainerDied","Data":"94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.609578 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9xsqt" event={"ID":"06164218-c618-4751-8f8c-d21b71f7bda5","Type":"ContainerDied","Data":"93027063b8fae75ab25320a11e611e0f6f0dbf6d14169c918dc2c29723a74600"} Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.609595 4750 scope.go:117] "RemoveContainer" containerID="94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.628086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.630854 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.130831333 +0000 UTC m=+251.473303731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.636937 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.641700 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wl7w6" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.647068 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-75zml" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.657586 4750 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-c9j6s container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.658103 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.659750 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9hnv9" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.689458 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.695097 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.700420 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.706609 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" podStartSLOduration=4.706585884 podStartE2EDuration="4.706585884s" podCreationTimestamp="2026-03-09 18:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:29.705043372 +0000 UTC m=+251.047515770" watchObservedRunningTime="2026-03-09 18:29:29.706585884 +0000 UTC m=+251.049058282" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.747426 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.749192 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.249173748 +0000 UTC m=+251.591646136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.799944 4750 scope.go:117] "RemoveContainer" containerID="94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4" Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.809588 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4\": container with ID starting with 94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4 not found: ID does not exist" containerID="94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.809656 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4"} err="failed to get container status \"94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4\": rpc error: code = NotFound desc = could not find container \"94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4\": container with ID starting with 94b0405a3dde93fde2d983c95286efe83fc4607037d522e25148667b23509ce4 not found: ID does not exist" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.849550 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.849960 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.349944293 +0000 UTC m=+251.692416691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.869012 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.952260 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:29 crc kubenswrapper[4750]: E0309 18:29:29.952659 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.452627499 +0000 UTC m=+251.795099887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:29 crc kubenswrapper[4750]: I0309 18:29:29.996964 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-44qdg" podStartSLOduration=181.996940691 podStartE2EDuration="3m1.996940691s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:29.910330023 +0000 UTC m=+251.252802441" watchObservedRunningTime="2026-03-09 18:29:29.996940691 +0000 UTC m=+251.339413089" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.018163 4750 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.053886 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:30 crc kubenswrapper[4750]: E0309 18:29:30.054242 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.554225067 +0000 UTC m=+251.896697465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.057489 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" podStartSLOduration=182.057474285 podStartE2EDuration="3m2.057474285s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:30.054817862 +0000 UTC m=+251.397290260" watchObservedRunningTime="2026-03-09 18:29:30.057474285 +0000 UTC m=+251.399946683" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.058163 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bv92v" podStartSLOduration=182.058158893 podStartE2EDuration="3m2.058158893s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:30.003199281 +0000 UTC m=+251.345671679" watchObservedRunningTime="2026-03-09 18:29:30.058158893 +0000 UTC m=+251.400631291" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.070873 4750 ???:1] "http: TLS handshake error from 192.168.126.11:35850: no serving certificate available for the kubelet" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.097617 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5c9r" podStartSLOduration=181.097594621 podStartE2EDuration="3m1.097594621s" podCreationTimestamp="2026-03-09 18:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:30.096192453 +0000 UTC m=+251.438664841" watchObservedRunningTime="2026-03-09 18:29:30.097594621 +0000 UTC m=+251.440067019" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.155308 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:30 crc kubenswrapper[4750]: E0309 18:29:30.155779 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.655751882 +0000 UTC m=+251.998224280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.214950 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:30 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:30 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:30 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.215032 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.259291 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:30 crc kubenswrapper[4750]: E0309 18:29:30.261004 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.760970447 +0000 UTC m=+252.103442855 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.297355 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9xsqt"] Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.308908 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9xsqt"] Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.361168 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c9b86b6b9-ndztb"] Mar 09 18:29:30 crc kubenswrapper[4750]: E0309 18:29:30.361407 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06164218-c618-4751-8f8c-d21b71f7bda5" containerName="controller-manager" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.361420 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="06164218-c618-4751-8f8c-d21b71f7bda5" containerName="controller-manager" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.361559 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="06164218-c618-4751-8f8c-d21b71f7bda5" containerName="controller-manager" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.361973 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.363449 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:30 crc kubenswrapper[4750]: E0309 18:29:30.363896 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.863877871 +0000 UTC m=+252.206350269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.382542 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.382884 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.383031 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.383154 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.383352 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.383462 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.412783 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c9b86b6b9-ndztb"] Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.437125 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.465498 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/538db02a-ecdb-4bc3-ba35-c972075f94c0-serving-cert\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.472982 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:30 crc kubenswrapper[4750]: E0309 18:29:30.480587 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 18:29:30.98055323 +0000 UTC m=+252.323025628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qmbbg" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.480935 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-proxy-ca-bundles\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.480993 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-config\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.481087 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-client-ca\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.481160 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttzm8\" (UniqueName: \"kubernetes.io/projected/538db02a-ecdb-4bc3-ba35-c972075f94c0-kube-api-access-ttzm8\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.584895 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.585405 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-proxy-ca-bundles\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.585493 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-config\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.585579 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-client-ca\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.586015 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttzm8\" (UniqueName: \"kubernetes.io/projected/538db02a-ecdb-4bc3-ba35-c972075f94c0-kube-api-access-ttzm8\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.586122 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/538db02a-ecdb-4bc3-ba35-c972075f94c0-serving-cert\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.589246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-config\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: E0309 18:29:30.589369 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 18:29:31.089347314 +0000 UTC m=+252.431819712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.590772 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-proxy-ca-bundles\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.591737 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-client-ca\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.605091 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/538db02a-ecdb-4bc3-ba35-c972075f94c0-serving-cert\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.631584 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttzm8\" (UniqueName: \"kubernetes.io/projected/538db02a-ecdb-4bc3-ba35-c972075f94c0-kube-api-access-ttzm8\") pod \"controller-manager-c9b86b6b9-ndztb\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.645466 4750 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-09T18:29:30.018195232Z","Handler":null,"Name":""} Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.647808 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" event={"ID":"90b2ac21-294a-40b6-b91e-8b189e5cca78","Type":"ContainerStarted","Data":"3348a4e61f68dd253b9ec2f892adb043e786344926b897c19f4153a51928da18"} Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.655518 4750 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.655585 4750 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.689813 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.696514 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.705857 4750 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.705936 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.824331 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rj4zw"] Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.836140 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.849120 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.849261 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rj4zw"] Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.860682 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:30 crc kubenswrapper[4750]: I0309 18:29:30.922107 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qmbbg\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.003352 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.003606 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-utilities\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.003670 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-catalog-content\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.003695 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-878gv\" (UniqueName: \"kubernetes.io/projected/41608683-d508-4374-a570-0e93ca339ffb-kube-api-access-878gv\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.027287 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.027843 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zvzmg"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.029005 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.042612 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.051910 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvzmg"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.109255 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.109574 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-utilities\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.109670 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-catalog-content\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.109713 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-catalog-content\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.109738 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btsjs\" (UniqueName: \"kubernetes.io/projected/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-kube-api-access-btsjs\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.109758 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-878gv\" (UniqueName: \"kubernetes.io/projected/41608683-d508-4374-a570-0e93ca339ffb-kube-api-access-878gv\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.109783 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-utilities\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.110322 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-utilities\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.110587 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-catalog-content\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.154678 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-878gv\" (UniqueName: \"kubernetes.io/projected/41608683-d508-4374-a570-0e93ca339ffb-kube-api-access-878gv\") pod \"community-operators-rj4zw\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.184412 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.206582 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:31 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:31 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:31 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.206674 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.211305 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-catalog-content\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.211378 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btsjs\" (UniqueName: \"kubernetes.io/projected/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-kube-api-access-btsjs\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.211407 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-utilities\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.212008 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-utilities\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.212329 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-catalog-content\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.241123 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.241951 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-86fld"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.252933 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.265721 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btsjs\" (UniqueName: \"kubernetes.io/projected/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-kube-api-access-btsjs\") pod \"certified-operators-zvzmg\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.284167 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86fld"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.314274 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ns2k\" (UniqueName: \"kubernetes.io/projected/3418a0da-42b0-4ac4-8957-87ed04209742-kube-api-access-2ns2k\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.314358 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-utilities\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.314405 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-catalog-content\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.417678 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-catalog-content\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.436705 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ns2k\" (UniqueName: \"kubernetes.io/projected/3418a0da-42b0-4ac4-8957-87ed04209742-kube-api-access-2ns2k\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.436813 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-utilities\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.437431 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-utilities\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.418233 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-catalog-content\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.457701 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06164218-c618-4751-8f8c-d21b71f7bda5" path="/var/lib/kubelet/pods/06164218-c618-4751-8f8c-d21b71f7bda5/volumes" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.459239 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.461436 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-wl7w6"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.461574 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tdsx8" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.462412 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dk5bm"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.464244 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk5bm"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.464318 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.466698 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.499247 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ns2k\" (UniqueName: \"kubernetes.io/projected/3418a0da-42b0-4ac4-8957-87ed04209742-kube-api-access-2ns2k\") pod \"community-operators-86fld\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.594118 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.644029 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-catalog-content\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.644117 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-utilities\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.644143 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c66jx\" (UniqueName: \"kubernetes.io/projected/7c78295a-888f-4690-b970-0e1557de4ad5-kube-api-access-c66jx\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.651430 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c9b86b6b9-ndztb"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.751744 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c66jx\" (UniqueName: \"kubernetes.io/projected/7c78295a-888f-4690-b970-0e1557de4ad5-kube-api-access-c66jx\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.751789 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-utilities\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.751855 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-catalog-content\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.752309 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-catalog-content\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.752950 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-utilities\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.792041 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c66jx\" (UniqueName: \"kubernetes.io/projected/7c78295a-888f-4690-b970-0e1557de4ad5-kube-api-access-c66jx\") pod \"certified-operators-dk5bm\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.805018 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" event={"ID":"cfb68d8f-61a3-40d7-8e89-44084f690dc2","Type":"ContainerStarted","Data":"97c8cdc0d0659692ef72929628e5b4a6112593c6958e3d7bc21100701d2cf19a"} Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.805088 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" event={"ID":"cfb68d8f-61a3-40d7-8e89-44084f690dc2","Type":"ContainerStarted","Data":"1dea19db87ca3decef40658f907fe3e5121966c0078f475fd7f39ac96337d2c4"} Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.843602 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1181187e83601fd2873c40ae4ed580e64bed1789900b5496624f778697f94ac4"} Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.854394 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rj4zw"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.868125 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" podStartSLOduration=183.868096287 podStartE2EDuration="3m3.868096287s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:31.862499284 +0000 UTC m=+253.204971682" watchObservedRunningTime="2026-03-09 18:29:31.868096287 +0000 UTC m=+253.210568685" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.876340 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qmbbg"] Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.898963 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:29:31 crc kubenswrapper[4750]: W0309 18:29:31.935064 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41608683_d508_4374_a570_0e93ca339ffb.slice/crio-2b22137fe94cb982150ca33e9853cc68d6d5b78e342e9d5c86cdac63751dbcf3 WatchSource:0}: Error finding container 2b22137fe94cb982150ca33e9853cc68d6d5b78e342e9d5c86cdac63751dbcf3: Status 404 returned error can't find the container with id 2b22137fe94cb982150ca33e9853cc68d6d5b78e342e9d5c86cdac63751dbcf3 Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.947177 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" event={"ID":"90b2ac21-294a-40b6-b91e-8b189e5cca78","Type":"ContainerStarted","Data":"5e832d86bdb24dbefbb6b8d01c100bba6db7ca4e72adb6c93b35a1c4f535b8cd"} Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.974588 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" event={"ID":"6a32a985-7c9a-4737-86db-d46edf0d0565","Type":"ContainerStarted","Data":"f58da1dd75d1a27240dc843ae8d69636b7eb6736a4beb5971a3c885395cc3b03"} Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.982975 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.984114 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.985250 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" event={"ID":"538db02a-ecdb-4bc3-ba35-c972075f94c0","Type":"ContainerStarted","Data":"ab9e1e64e42feec378b7bf48c4901f3a289913018ceddd454943d298ac0a4896"} Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.990706 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d46e72c785c5a520b76ed95518e2956c18a0e0617735db8296a712d8efbe330b"} Mar 09 18:29:31 crc kubenswrapper[4750]: I0309 18:29:31.998892 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-5pntt" podStartSLOduration=12.998859641 podStartE2EDuration="12.998859641s" podCreationTimestamp="2026-03-09 18:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:31.981530127 +0000 UTC m=+253.324002525" watchObservedRunningTime="2026-03-09 18:29:31.998859641 +0000 UTC m=+253.341332039" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.002924 4750 patch_prober.go:28] interesting pod/console-f9d7485db-bglg2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.003023 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bglg2" podUID="863600af-6300-46d6-875a-85b0622c8f50" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.006609 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"efddd2c9d52fd9e156041a61f37b29c71289800a600e335f438dbf32e84023fe"} Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.010549 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvzmg"] Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.018694 4750 generic.go:334] "Generic (PLEG): container finished" podID="d0ecde8f-d9fa-45e5-a508-981772436b4f" containerID="1aa681e696281cc67b93af1e699c7c67d083d7656749096e07580423b92d0786" exitCode=0 Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.021146 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" event={"ID":"d0ecde8f-d9fa-45e5-a508-981772436b4f","Type":"ContainerDied","Data":"1aa681e696281cc67b93af1e699c7c67d083d7656749096e07580423b92d0786"} Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.194886 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.213762 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:32 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:32 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:32 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.213839 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.314711 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86fld"] Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.358110 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-scg4c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.358163 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-scg4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.358188 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-scg4c" podUID="43705153-a050-4d95-ba46-792661ace7ae" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.358237 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-scg4c" podUID="43705153-a050-4d95-ba46-792661ace7ae" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.493608 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk5bm"] Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.810842 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jk4f9"] Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.824856 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.827783 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.829498 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk4f9"] Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.900130 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x27vz\" (UniqueName: \"kubernetes.io/projected/2de1257e-0c9d-486f-92da-2e0525941f7e-kube-api-access-x27vz\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.900194 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-catalog-content\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:32 crc kubenswrapper[4750]: I0309 18:29:32.900238 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-utilities\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.001983 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x27vz\" (UniqueName: \"kubernetes.io/projected/2de1257e-0c9d-486f-92da-2e0525941f7e-kube-api-access-x27vz\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.002080 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-catalog-content\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.002127 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-utilities\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.003253 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-catalog-content\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.003288 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-utilities\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.033447 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d243a0ef27cdc579f2c81c93ae93f6895ef7925210bd081fb2952cffab757338"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.035314 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.038031 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x27vz\" (UniqueName: \"kubernetes.io/projected/2de1257e-0c9d-486f-92da-2e0525941f7e-kube-api-access-x27vz\") pod \"redhat-marketplace-jk4f9\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.041421 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"dc860bf37b3afa587f0de21f1cae943247d078cfff5e4fc090cf98a591dae14d"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.050762 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" event={"ID":"74eac05b-f613-45a9-b8fa-1eb83e8c04b4","Type":"ContainerStarted","Data":"d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.050817 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" event={"ID":"74eac05b-f613-45a9-b8fa-1eb83e8c04b4","Type":"ContainerStarted","Data":"ed312d0f39a71a1a8cfd66600a36b9b3443f0c153fcb647cdae15d549afe1416"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.051841 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.068144 4750 generic.go:334] "Generic (PLEG): container finished" podID="3418a0da-42b0-4ac4-8957-87ed04209742" containerID="6bf37f8d34bfc259388ee223d5327a75cb31a178b9fa3eef0c76d2032b060020" exitCode=0 Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.068235 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86fld" event={"ID":"3418a0da-42b0-4ac4-8957-87ed04209742","Type":"ContainerDied","Data":"6bf37f8d34bfc259388ee223d5327a75cb31a178b9fa3eef0c76d2032b060020"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.068269 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86fld" event={"ID":"3418a0da-42b0-4ac4-8957-87ed04209742","Type":"ContainerStarted","Data":"ae94e61e466a995bad1ca251a6e3728b05948efb2bac95213dcc4da95ecea0cd"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.091131 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" event={"ID":"6a32a985-7c9a-4737-86db-d46edf0d0565","Type":"ContainerStarted","Data":"662487cb1824e28a32a551dee4cd364128e0a24b2fb5185b54dd153ad8202ac4"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.091222 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wl7w6" event={"ID":"6a32a985-7c9a-4737-86db-d46edf0d0565","Type":"ContainerStarted","Data":"59fa40494f5de8d086cbd5d73db8c0e496f7e8060c55a594503b2967837bf808"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.112083 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" event={"ID":"538db02a-ecdb-4bc3-ba35-c972075f94c0","Type":"ContainerStarted","Data":"56c42401ee9bb635b4682f342e5a17e34ba09f55613dfd770639025a58e9f113"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.113788 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.142102 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.142148 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.142878 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.154939 4750 generic.go:334] "Generic (PLEG): container finished" podID="7c78295a-888f-4690-b970-0e1557de4ad5" containerID="a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2" exitCode=0 Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.155077 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk5bm" event={"ID":"7c78295a-888f-4690-b970-0e1557de4ad5","Type":"ContainerDied","Data":"a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.155118 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk5bm" event={"ID":"7c78295a-888f-4690-b970-0e1557de4ad5","Type":"ContainerStarted","Data":"2ebb33a38039631a51c4c042920f0ebe9aebc375df85a06ec1041d607cb4a2d7"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.162887 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.166706 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" podStartSLOduration=185.166683452 podStartE2EDuration="3m5.166683452s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:33.123671747 +0000 UTC m=+254.466144145" watchObservedRunningTime="2026-03-09 18:29:33.166683452 +0000 UTC m=+254.509155850" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.202394 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.205300 4750 generic.go:334] "Generic (PLEG): container finished" podID="41608683-d508-4374-a570-0e93ca339ffb" containerID="fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7" exitCode=0 Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.205485 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rj4zw" event={"ID":"41608683-d508-4374-a570-0e93ca339ffb","Type":"ContainerDied","Data":"fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.205526 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rj4zw" event={"ID":"41608683-d508-4374-a570-0e93ca339ffb","Type":"ContainerStarted","Data":"2b22137fe94cb982150ca33e9853cc68d6d5b78e342e9d5c86cdac63751dbcf3"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.214934 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:33 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:33 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:33 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.214991 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.250947 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v879l"] Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.253075 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.255418 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"12cb068855c9e840480e44ebb09ea19d1da0e791e19c5cc960c5dc45e2c0d53e"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.286682 4750 generic.go:334] "Generic (PLEG): container finished" podID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerID="0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11" exitCode=0 Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.294983 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvzmg" event={"ID":"ab00ccba-a824-42a2-a9fc-43139bcf9a0e","Type":"ContainerDied","Data":"0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.295066 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvzmg" event={"ID":"ab00ccba-a824-42a2-a9fc-43139bcf9a0e","Type":"ContainerStarted","Data":"da7388fecb1b148c9444914252ca05d16e962fa136b3f760811b5a5bfe63db3d"} Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.319763 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v879l"] Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.320073 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" podStartSLOduration=8.320052564 podStartE2EDuration="8.320052564s" podCreationTimestamp="2026-03-09 18:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:33.305829716 +0000 UTC m=+254.648302114" watchObservedRunningTime="2026-03-09 18:29:33.320052564 +0000 UTC m=+254.662524962" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.423605 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k5lr\" (UniqueName: \"kubernetes.io/projected/b880ae0c-0f6f-487b-8927-77437bb8a638-kube-api-access-9k5lr\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.423772 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-utilities\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.424587 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-catalog-content\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.434254 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-wl7w6" podStartSLOduration=185.434227695 podStartE2EDuration="3m5.434227695s" podCreationTimestamp="2026-03-09 18:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:33.389297166 +0000 UTC m=+254.731769574" watchObservedRunningTime="2026-03-09 18:29:33.434227695 +0000 UTC m=+254.776700093" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.527611 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-catalog-content\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.528215 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k5lr\" (UniqueName: \"kubernetes.io/projected/b880ae0c-0f6f-487b-8927-77437bb8a638-kube-api-access-9k5lr\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.528256 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-utilities\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.528376 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-catalog-content\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.528791 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-utilities\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.560456 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k5lr\" (UniqueName: \"kubernetes.io/projected/b880ae0c-0f6f-487b-8927-77437bb8a638-kube-api-access-9k5lr\") pod \"redhat-marketplace-v879l\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.619704 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.664614 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.664714 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.776850 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.910590 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk4f9"] Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.943701 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcmd9\" (UniqueName: \"kubernetes.io/projected/d0ecde8f-d9fa-45e5-a508-981772436b4f-kube-api-access-dcmd9\") pod \"d0ecde8f-d9fa-45e5-a508-981772436b4f\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.943840 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ecde8f-d9fa-45e5-a508-981772436b4f-config-volume\") pod \"d0ecde8f-d9fa-45e5-a508-981772436b4f\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.943936 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ecde8f-d9fa-45e5-a508-981772436b4f-secret-volume\") pod \"d0ecde8f-d9fa-45e5-a508-981772436b4f\" (UID: \"d0ecde8f-d9fa-45e5-a508-981772436b4f\") " Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.947752 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0ecde8f-d9fa-45e5-a508-981772436b4f-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0ecde8f-d9fa-45e5-a508-981772436b4f" (UID: "d0ecde8f-d9fa-45e5-a508-981772436b4f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.953008 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ecde8f-d9fa-45e5-a508-981772436b4f-kube-api-access-dcmd9" (OuterVolumeSpecName: "kube-api-access-dcmd9") pod "d0ecde8f-d9fa-45e5-a508-981772436b4f" (UID: "d0ecde8f-d9fa-45e5-a508-981772436b4f"). InnerVolumeSpecName "kube-api-access-dcmd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:29:33 crc kubenswrapper[4750]: I0309 18:29:33.962046 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ecde8f-d9fa-45e5-a508-981772436b4f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0ecde8f-d9fa-45e5-a508-981772436b4f" (UID: "d0ecde8f-d9fa-45e5-a508-981772436b4f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.039328 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lxwd2"] Mar 09 18:29:34 crc kubenswrapper[4750]: E0309 18:29:34.039799 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ecde8f-d9fa-45e5-a508-981772436b4f" containerName="collect-profiles" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.039816 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ecde8f-d9fa-45e5-a508-981772436b4f" containerName="collect-profiles" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.039974 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ecde8f-d9fa-45e5-a508-981772436b4f" containerName="collect-profiles" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.041097 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.044539 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.045344 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw7mj\" (UniqueName: \"kubernetes.io/projected/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-kube-api-access-lw7mj\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.045407 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-catalog-content\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.045446 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-utilities\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.045484 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcmd9\" (UniqueName: \"kubernetes.io/projected/d0ecde8f-d9fa-45e5-a508-981772436b4f-kube-api-access-dcmd9\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.045494 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0ecde8f-d9fa-45e5-a508-981772436b4f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.045503 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0ecde8f-d9fa-45e5-a508-981772436b4f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.055951 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxwd2"] Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.069609 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.079150 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.088885 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.094867 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.094983 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.148872 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-utilities\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.148918 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw7mj\" (UniqueName: \"kubernetes.io/projected/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-kube-api-access-lw7mj\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.148970 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-catalog-content\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.149534 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-catalog-content\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.149821 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-utilities\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.202584 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:34 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:34 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:34 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.202653 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.204536 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw7mj\" (UniqueName: \"kubernetes.io/projected/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-kube-api-access-lw7mj\") pod \"redhat-operators-lxwd2\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.251797 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.251872 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.264761 4750 patch_prober.go:28] interesting pod/apiserver-76f77b778f-pctzl container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]log ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]etcd ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/generic-apiserver-start-informers ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/max-in-flight-filter ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 09 18:29:34 crc kubenswrapper[4750]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 09 18:29:34 crc kubenswrapper[4750]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/project.openshift.io-projectcache ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/openshift.io-startinformers ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 09 18:29:34 crc kubenswrapper[4750]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 09 18:29:34 crc kubenswrapper[4750]: livez check failed Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.264845 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" podUID="cfb68d8f-61a3-40d7-8e89-44084f690dc2" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.356999 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.357354 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.357945 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.375377 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.375400 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr" event={"ID":"d0ecde8f-d9fa-45e5-a508-981772436b4f","Type":"ContainerDied","Data":"5f2cd1804ad59b35c14c42a91bba3c8b769ee5bb20bb1af6354d213391070e21"} Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.375725 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f2cd1804ad59b35c14c42a91bba3c8b769ee5bb20bb1af6354d213391070e21" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.379201 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.394343 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.394410 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk4f9" event={"ID":"2de1257e-0c9d-486f-92da-2e0525941f7e","Type":"ContainerStarted","Data":"7505bda44a51b5cb2d2d01982524f8c840a7bf927107e2587fb6aed9052a5ee2"} Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.405838 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j765j" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.428275 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-76nqh"] Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.432514 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.446311 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-76nqh"] Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.472748 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v879l"] Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.497165 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.560204 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn6pd\" (UniqueName: \"kubernetes.io/projected/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-kube-api-access-dn6pd\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.560856 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-utilities\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.561119 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-catalog-content\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: W0309 18:29:34.569739 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb880ae0c_0f6f_487b_8927_77437bb8a638.slice/crio-07ce79bd8479ba3c148f0eba9a93d1882bcc6e5310ce64049d3f04dc917f27ef WatchSource:0}: Error finding container 07ce79bd8479ba3c148f0eba9a93d1882bcc6e5310ce64049d3f04dc917f27ef: Status 404 returned error can't find the container with id 07ce79bd8479ba3c148f0eba9a93d1882bcc6e5310ce64049d3f04dc917f27ef Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.664585 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-utilities\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.664676 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-catalog-content\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.664730 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn6pd\" (UniqueName: \"kubernetes.io/projected/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-kube-api-access-dn6pd\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.665602 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-utilities\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.665912 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-catalog-content\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.720624 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn6pd\" (UniqueName: \"kubernetes.io/projected/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-kube-api-access-dn6pd\") pod \"redhat-operators-76nqh\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.832533 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.934974 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.936160 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.952421 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.952647 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 09 18:29:34 crc kubenswrapper[4750]: I0309 18:29:34.960732 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.075604 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.075673 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.180567 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.180615 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.181048 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.198982 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:35 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:35 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:35 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.199039 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.228555 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.260105 4750 ???:1] "http: TLS handshake error from 192.168.126.11:35854: no serving certificate available for the kubelet" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.342963 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.503778 4750 generic.go:334] "Generic (PLEG): container finished" podID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerID="21d7ac9d0e4961907d815247e020c5032d2b2aa88b34353240f94a9753b074bc" exitCode=0 Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.506523 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v879l" event={"ID":"b880ae0c-0f6f-487b-8927-77437bb8a638","Type":"ContainerDied","Data":"21d7ac9d0e4961907d815247e020c5032d2b2aa88b34353240f94a9753b074bc"} Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.507316 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v879l" event={"ID":"b880ae0c-0f6f-487b-8927-77437bb8a638","Type":"ContainerStarted","Data":"07ce79bd8479ba3c148f0eba9a93d1882bcc6e5310ce64049d3f04dc917f27ef"} Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.522272 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxwd2"] Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.567949 4750 generic.go:334] "Generic (PLEG): container finished" podID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerID="211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0" exitCode=0 Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.569424 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk4f9" event={"ID":"2de1257e-0c9d-486f-92da-2e0525941f7e","Type":"ContainerDied","Data":"211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0"} Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.642976 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 09 18:29:35 crc kubenswrapper[4750]: I0309 18:29:35.904117 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-76nqh"] Mar 09 18:29:35 crc kubenswrapper[4750]: W0309 18:29:35.963254 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c35f63c_cbc8_4d2b_b971_3e18a8222bc7.slice/crio-f6413534428f3d4968e7e1b5177af59980b14d8573513ff1322749b2c048fcb6 WatchSource:0}: Error finding container f6413534428f3d4968e7e1b5177af59980b14d8573513ff1322749b2c048fcb6: Status 404 returned error can't find the container with id f6413534428f3d4968e7e1b5177af59980b14d8573513ff1322749b2c048fcb6 Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.100905 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.212838 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:36 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:36 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:36 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.213370 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.599981 4750 generic.go:334] "Generic (PLEG): container finished" podID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerID="eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092" exitCode=0 Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.600064 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxwd2" event={"ID":"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082","Type":"ContainerDied","Data":"eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092"} Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.600097 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxwd2" event={"ID":"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082","Type":"ContainerStarted","Data":"9af23198bc0b1f7e1d2c46fa1368dcc4c78a418be0fe63e0fc2db838c92b33fe"} Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.608368 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7","Type":"ContainerStarted","Data":"cad32e7031f52b68e0e427fb6c495cfc79aa52d6a9c64e7191621492638ab0ca"} Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.668287 4750 generic.go:334] "Generic (PLEG): container finished" podID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerID="e423261a174b2b891ba24702f2e32a44290a58ae99ca00baca80127fcc89780f" exitCode=0 Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.668395 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76nqh" event={"ID":"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7","Type":"ContainerDied","Data":"e423261a174b2b891ba24702f2e32a44290a58ae99ca00baca80127fcc89780f"} Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.668511 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76nqh" event={"ID":"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7","Type":"ContainerStarted","Data":"f6413534428f3d4968e7e1b5177af59980b14d8573513ff1322749b2c048fcb6"} Mar 09 18:29:36 crc kubenswrapper[4750]: I0309 18:29:36.695326 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e9e67ec-16bf-46e3-9d5f-07848b4af04d","Type":"ContainerStarted","Data":"f6e71b4bfd951a268ed812f693f8391e24fb105ddefcf64852a63eb13931cddb"} Mar 09 18:29:37 crc kubenswrapper[4750]: I0309 18:29:37.200493 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:37 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:37 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:37 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:37 crc kubenswrapper[4750]: I0309 18:29:37.201100 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:37 crc kubenswrapper[4750]: I0309 18:29:37.731085 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7","Type":"ContainerStarted","Data":"5614e781c049ba687e59e22bccee72cdf332f503bb97c430e0ad5122fdad3136"} Mar 09 18:29:37 crc kubenswrapper[4750]: I0309 18:29:37.744235 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e9e67ec-16bf-46e3-9d5f-07848b4af04d","Type":"ContainerStarted","Data":"fd53c69b2e9ed3cbcd82ebe7174d33492049c7a93f0216256485f76643098456"} Mar 09 18:29:37 crc kubenswrapper[4750]: I0309 18:29:37.751513 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.751468092 podStartE2EDuration="3.751468092s" podCreationTimestamp="2026-03-09 18:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:37.748155572 +0000 UTC m=+259.090627970" watchObservedRunningTime="2026-03-09 18:29:37.751468092 +0000 UTC m=+259.093940490" Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.198446 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:38 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:38 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:38 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.199131 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.577245 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-qjrwx" Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.603768 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.603748529 podStartE2EDuration="4.603748529s" podCreationTimestamp="2026-03-09 18:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:29:37.775405247 +0000 UTC m=+259.117877645" watchObservedRunningTime="2026-03-09 18:29:38.603748529 +0000 UTC m=+259.946220927" Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.685162 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.699868 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pctzl" Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.786041 4750 generic.go:334] "Generic (PLEG): container finished" podID="3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7" containerID="5614e781c049ba687e59e22bccee72cdf332f503bb97c430e0ad5122fdad3136" exitCode=0 Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.786137 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7","Type":"ContainerDied","Data":"5614e781c049ba687e59e22bccee72cdf332f503bb97c430e0ad5122fdad3136"} Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.808858 4750 generic.go:334] "Generic (PLEG): container finished" podID="9e9e67ec-16bf-46e3-9d5f-07848b4af04d" containerID="fd53c69b2e9ed3cbcd82ebe7174d33492049c7a93f0216256485f76643098456" exitCode=0 Mar 09 18:29:38 crc kubenswrapper[4750]: I0309 18:29:38.809045 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e9e67ec-16bf-46e3-9d5f-07848b4af04d","Type":"ContainerDied","Data":"fd53c69b2e9ed3cbcd82ebe7174d33492049c7a93f0216256485f76643098456"} Mar 09 18:29:39 crc kubenswrapper[4750]: I0309 18:29:39.198578 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:39 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:39 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:39 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:39 crc kubenswrapper[4750]: I0309 18:29:39.198651 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:40 crc kubenswrapper[4750]: I0309 18:29:40.219016 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:40 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:40 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:40 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:40 crc kubenswrapper[4750]: I0309 18:29:40.219085 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:41 crc kubenswrapper[4750]: I0309 18:29:41.202506 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:41 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:41 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:41 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:41 crc kubenswrapper[4750]: I0309 18:29:41.203487 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:41 crc kubenswrapper[4750]: I0309 18:29:41.691261 4750 ???:1] "http: TLS handshake error from 192.168.126.11:60316: no serving certificate available for the kubelet" Mar 09 18:29:41 crc kubenswrapper[4750]: I0309 18:29:41.981234 4750 patch_prober.go:28] interesting pod/console-f9d7485db-bglg2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 09 18:29:41 crc kubenswrapper[4750]: I0309 18:29:41.981768 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bglg2" podUID="863600af-6300-46d6-875a-85b0622c8f50" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 09 18:29:42 crc kubenswrapper[4750]: I0309 18:29:42.198442 4750 patch_prober.go:28] interesting pod/router-default-5444994796-cwxb8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 18:29:42 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Mar 09 18:29:42 crc kubenswrapper[4750]: [+]process-running ok Mar 09 18:29:42 crc kubenswrapper[4750]: healthz check failed Mar 09 18:29:42 crc kubenswrapper[4750]: I0309 18:29:42.198518 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cwxb8" podUID="3789c695-4f63-465e-9ccb-90ee29189e38" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:29:42 crc kubenswrapper[4750]: I0309 18:29:42.355193 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-scg4c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Mar 09 18:29:42 crc kubenswrapper[4750]: I0309 18:29:42.355277 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-scg4c" podUID="43705153-a050-4d95-ba46-792661ace7ae" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Mar 09 18:29:42 crc kubenswrapper[4750]: I0309 18:29:42.355382 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-scg4c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Mar 09 18:29:42 crc kubenswrapper[4750]: I0309 18:29:42.355444 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-scg4c" podUID="43705153-a050-4d95-ba46-792661ace7ae" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Mar 09 18:29:43 crc kubenswrapper[4750]: I0309 18:29:43.199121 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:43 crc kubenswrapper[4750]: I0309 18:29:43.202036 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cwxb8" Mar 09 18:29:44 crc kubenswrapper[4750]: I0309 18:29:44.488420 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c9b86b6b9-ndztb"] Mar 09 18:29:44 crc kubenswrapper[4750]: I0309 18:29:44.488851 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" podUID="538db02a-ecdb-4bc3-ba35-c972075f94c0" containerName="controller-manager" containerID="cri-o://56c42401ee9bb635b4682f342e5a17e34ba09f55613dfd770639025a58e9f113" gracePeriod=30 Mar 09 18:29:44 crc kubenswrapper[4750]: I0309 18:29:44.530006 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79"] Mar 09 18:29:44 crc kubenswrapper[4750]: I0309 18:29:44.530690 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" podUID="178203e6-eb24-4f4f-b794-b11b496778e4" containerName="route-controller-manager" containerID="cri-o://0ca0155e7ea9f0108cf9a825a6a8d9b2a6513ff48fa06e8ff5cde596fe50e039" gracePeriod=30 Mar 09 18:29:44 crc kubenswrapper[4750]: I0309 18:29:44.913193 4750 generic.go:334] "Generic (PLEG): container finished" podID="178203e6-eb24-4f4f-b794-b11b496778e4" containerID="0ca0155e7ea9f0108cf9a825a6a8d9b2a6513ff48fa06e8ff5cde596fe50e039" exitCode=0 Mar 09 18:29:44 crc kubenswrapper[4750]: I0309 18:29:44.913246 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" event={"ID":"178203e6-eb24-4f4f-b794-b11b496778e4","Type":"ContainerDied","Data":"0ca0155e7ea9f0108cf9a825a6a8d9b2a6513ff48fa06e8ff5cde596fe50e039"} Mar 09 18:29:45 crc kubenswrapper[4750]: I0309 18:29:45.536056 4750 ???:1] "http: TLS handshake error from 192.168.126.11:60324: no serving certificate available for the kubelet" Mar 09 18:29:45 crc kubenswrapper[4750]: I0309 18:29:45.930305 4750 generic.go:334] "Generic (PLEG): container finished" podID="538db02a-ecdb-4bc3-ba35-c972075f94c0" containerID="56c42401ee9bb635b4682f342e5a17e34ba09f55613dfd770639025a58e9f113" exitCode=0 Mar 09 18:29:45 crc kubenswrapper[4750]: I0309 18:29:45.930376 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" event={"ID":"538db02a-ecdb-4bc3-ba35-c972075f94c0","Type":"ContainerDied","Data":"56c42401ee9bb635b4682f342e5a17e34ba09f55613dfd770639025a58e9f113"} Mar 09 18:29:47 crc kubenswrapper[4750]: I0309 18:29:47.793811 4750 patch_prober.go:28] interesting pod/route-controller-manager-66dc67c9bf-vzb79 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" start-of-body= Mar 09 18:29:47 crc kubenswrapper[4750]: I0309 18:29:47.794514 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" podUID="178203e6-eb24-4f4f-b794-b11b496778e4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.811287 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.819870 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.864547 4750 patch_prober.go:28] interesting pod/controller-manager-c9b86b6b9-ndztb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.46:8443/healthz\": dial tcp 10.217.0.46:8443: connect: connection refused" start-of-body= Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.864648 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" podUID="538db02a-ecdb-4bc3-ba35-c972075f94c0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.46:8443/healthz\": dial tcp 10.217.0.46:8443: connect: connection refused" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.882809 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kube-api-access\") pod \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.882915 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kubelet-dir\") pod \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.882936 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kubelet-dir\") pod \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\" (UID: \"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7\") " Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.883026 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kube-api-access\") pod \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\" (UID: \"9e9e67ec-16bf-46e3-9d5f-07848b4af04d\") " Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.883128 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e9e67ec-16bf-46e3-9d5f-07848b4af04d" (UID: "9e9e67ec-16bf-46e3-9d5f-07848b4af04d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.883162 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7" (UID: "3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.883331 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.883349 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.903459 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e9e67ec-16bf-46e3-9d5f-07848b4af04d" (UID: "9e9e67ec-16bf-46e3-9d5f-07848b4af04d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.915729 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7" (UID: "3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.978261 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7","Type":"ContainerDied","Data":"cad32e7031f52b68e0e427fb6c495cfc79aa52d6a9c64e7191621492638ab0ca"} Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.978332 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cad32e7031f52b68e0e427fb6c495cfc79aa52d6a9c64e7191621492638ab0ca" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.978282 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.980521 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e9e67ec-16bf-46e3-9d5f-07848b4af04d","Type":"ContainerDied","Data":"f6e71b4bfd951a268ed812f693f8391e24fb105ddefcf64852a63eb13931cddb"} Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.980570 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6e71b4bfd951a268ed812f693f8391e24fb105ddefcf64852a63eb13931cddb" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.980663 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.984472 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9e67ec-16bf-46e3-9d5f-07848b4af04d-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:50 crc kubenswrapper[4750]: I0309 18:29:50.984499 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:51 crc kubenswrapper[4750]: I0309 18:29:51.213870 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:29:51 crc kubenswrapper[4750]: I0309 18:29:51.743995 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:29:51 crc kubenswrapper[4750]: I0309 18:29:51.744075 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:29:51 crc kubenswrapper[4750]: I0309 18:29:51.986856 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:51 crc kubenswrapper[4750]: I0309 18:29:51.993126 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:29:52 crc kubenswrapper[4750]: I0309 18:29:52.364266 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-scg4c" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.064930 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.142093 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttzm8\" (UniqueName: \"kubernetes.io/projected/538db02a-ecdb-4bc3-ba35-c972075f94c0-kube-api-access-ttzm8\") pod \"538db02a-ecdb-4bc3-ba35-c972075f94c0\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.142205 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-config\") pod \"538db02a-ecdb-4bc3-ba35-c972075f94c0\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.142244 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/538db02a-ecdb-4bc3-ba35-c972075f94c0-serving-cert\") pod \"538db02a-ecdb-4bc3-ba35-c972075f94c0\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.142290 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-proxy-ca-bundles\") pod \"538db02a-ecdb-4bc3-ba35-c972075f94c0\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.142322 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-client-ca\") pod \"538db02a-ecdb-4bc3-ba35-c972075f94c0\" (UID: \"538db02a-ecdb-4bc3-ba35-c972075f94c0\") " Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.143522 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-client-ca" (OuterVolumeSpecName: "client-ca") pod "538db02a-ecdb-4bc3-ba35-c972075f94c0" (UID: "538db02a-ecdb-4bc3-ba35-c972075f94c0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.144009 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "538db02a-ecdb-4bc3-ba35-c972075f94c0" (UID: "538db02a-ecdb-4bc3-ba35-c972075f94c0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.144180 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-config" (OuterVolumeSpecName: "config") pod "538db02a-ecdb-4bc3-ba35-c972075f94c0" (UID: "538db02a-ecdb-4bc3-ba35-c972075f94c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.154216 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/538db02a-ecdb-4bc3-ba35-c972075f94c0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "538db02a-ecdb-4bc3-ba35-c972075f94c0" (UID: "538db02a-ecdb-4bc3-ba35-c972075f94c0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.154342 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538db02a-ecdb-4bc3-ba35-c972075f94c0-kube-api-access-ttzm8" (OuterVolumeSpecName: "kube-api-access-ttzm8") pod "538db02a-ecdb-4bc3-ba35-c972075f94c0" (UID: "538db02a-ecdb-4bc3-ba35-c972075f94c0"). InnerVolumeSpecName "kube-api-access-ttzm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.244064 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.244108 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/538db02a-ecdb-4bc3-ba35-c972075f94c0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.244118 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.244131 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/538db02a-ecdb-4bc3-ba35-c972075f94c0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:54 crc kubenswrapper[4750]: I0309 18:29:54.244140 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttzm8\" (UniqueName: \"kubernetes.io/projected/538db02a-ecdb-4bc3-ba35-c972075f94c0-kube-api-access-ttzm8\") on node \"crc\" DevicePath \"\"" Mar 09 18:29:55 crc kubenswrapper[4750]: I0309 18:29:55.012996 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" event={"ID":"538db02a-ecdb-4bc3-ba35-c972075f94c0","Type":"ContainerDied","Data":"ab9e1e64e42feec378b7bf48c4901f3a289913018ceddd454943d298ac0a4896"} Mar 09 18:29:55 crc kubenswrapper[4750]: I0309 18:29:55.013064 4750 scope.go:117] "RemoveContainer" containerID="56c42401ee9bb635b4682f342e5a17e34ba09f55613dfd770639025a58e9f113" Mar 09 18:29:55 crc kubenswrapper[4750]: I0309 18:29:55.013070 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c9b86b6b9-ndztb" Mar 09 18:29:55 crc kubenswrapper[4750]: I0309 18:29:55.046034 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c9b86b6b9-ndztb"] Mar 09 18:29:55 crc kubenswrapper[4750]: I0309 18:29:55.049660 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-c9b86b6b9-ndztb"] Mar 09 18:29:55 crc kubenswrapper[4750]: I0309 18:29:55.381496 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538db02a-ecdb-4bc3-ba35-c972075f94c0" path="/var/lib/kubelet/pods/538db02a-ecdb-4bc3-ba35-c972075f94c0/volumes" Mar 09 18:29:56 crc kubenswrapper[4750]: E0309 18:29:56.637939 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 09 18:29:56 crc kubenswrapper[4750]: E0309 18:29:56.638157 4750 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 18:29:56 crc kubenswrapper[4750]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 09 18:29:56 crc kubenswrapper[4750]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fk6ng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29551348-g9vjb_openshift-infra(a08fcb79-42d8-47c2-940e-92c4ff4f2f61): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 09 18:29:56 crc kubenswrapper[4750]: > logger="UnhandledError" Mar 09 18:29:56 crc kubenswrapper[4750]: E0309 18:29:56.640590 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" podUID="a08fcb79-42d8-47c2-940e-92c4ff4f2f61" Mar 09 18:29:57 crc kubenswrapper[4750]: E0309 18:29:57.031924 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" podUID="a08fcb79-42d8-47c2-940e-92c4ff4f2f61" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.360939 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b874b4dd8-bhvkp"] Mar 09 18:29:57 crc kubenswrapper[4750]: E0309 18:29:57.361206 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7" containerName="pruner" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.361220 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7" containerName="pruner" Mar 09 18:29:57 crc kubenswrapper[4750]: E0309 18:29:57.361230 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538db02a-ecdb-4bc3-ba35-c972075f94c0" containerName="controller-manager" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.361236 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="538db02a-ecdb-4bc3-ba35-c972075f94c0" containerName="controller-manager" Mar 09 18:29:57 crc kubenswrapper[4750]: E0309 18:29:57.361244 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9e67ec-16bf-46e3-9d5f-07848b4af04d" containerName="pruner" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.361253 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9e67ec-16bf-46e3-9d5f-07848b4af04d" containerName="pruner" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.361355 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="538db02a-ecdb-4bc3-ba35-c972075f94c0" containerName="controller-manager" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.361366 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9e67ec-16bf-46e3-9d5f-07848b4af04d" containerName="pruner" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.361380 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3f8b45-aa5b-4121-92a3-4dcfb3784ec7" containerName="pruner" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.361996 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.366002 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.366033 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.366524 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.367055 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.367112 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.367494 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.373934 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.388763 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b874b4dd8-bhvkp"] Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.407738 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dswp\" (UniqueName: \"kubernetes.io/projected/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-kube-api-access-8dswp\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.407807 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-config\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.407838 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-serving-cert\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.407862 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-client-ca\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.407899 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-proxy-ca-bundles\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.509161 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-config\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.509246 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-serving-cert\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.509272 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-client-ca\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.509388 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-proxy-ca-bundles\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.509434 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dswp\" (UniqueName: \"kubernetes.io/projected/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-kube-api-access-8dswp\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.511996 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-proxy-ca-bundles\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.512183 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-client-ca\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.513250 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-config\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.518341 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-serving-cert\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.530244 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dswp\" (UniqueName: \"kubernetes.io/projected/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-kube-api-access-8dswp\") pod \"controller-manager-b874b4dd8-bhvkp\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:57 crc kubenswrapper[4750]: I0309 18:29:57.689416 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:29:58 crc kubenswrapper[4750]: I0309 18:29:58.794251 4750 patch_prober.go:28] interesting pod/route-controller-manager-66dc67c9bf-vzb79 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:29:58 crc kubenswrapper[4750]: I0309 18:29:58.794395 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" podUID="178203e6-eb24-4f4f-b794-b11b496778e4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.45:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.052862 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" event={"ID":"178203e6-eb24-4f4f-b794-b11b496778e4","Type":"ContainerDied","Data":"d697a49b3584135834944a4fad0af885c5515316a20be8c8fb691266dd57a860"} Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.052929 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d697a49b3584135834944a4fad0af885c5515316a20be8c8fb691266dd57a860" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.060479 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.143526 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551350-vmx4v"] Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.145661 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/178203e6-eb24-4f4f-b794-b11b496778e4-serving-cert\") pod \"178203e6-eb24-4f4f-b794-b11b496778e4\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.145727 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-client-ca\") pod \"178203e6-eb24-4f4f-b794-b11b496778e4\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.145767 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrt2s\" (UniqueName: \"kubernetes.io/projected/178203e6-eb24-4f4f-b794-b11b496778e4-kube-api-access-jrt2s\") pod \"178203e6-eb24-4f4f-b794-b11b496778e4\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " Mar 09 18:30:00 crc kubenswrapper[4750]: E0309 18:30:00.145815 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178203e6-eb24-4f4f-b794-b11b496778e4" containerName="route-controller-manager" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.145848 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="178203e6-eb24-4f4f-b794-b11b496778e4" containerName="route-controller-manager" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.145895 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-config\") pod \"178203e6-eb24-4f4f-b794-b11b496778e4\" (UID: \"178203e6-eb24-4f4f-b794-b11b496778e4\") " Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.146045 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="178203e6-eb24-4f4f-b794-b11b496778e4" containerName="route-controller-manager" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.146813 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551350-vmx4v" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.146803 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-client-ca" (OuterVolumeSpecName: "client-ca") pod "178203e6-eb24-4f4f-b794-b11b496778e4" (UID: "178203e6-eb24-4f4f-b794-b11b496778e4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.146826 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-config" (OuterVolumeSpecName: "config") pod "178203e6-eb24-4f4f-b794-b11b496778e4" (UID: "178203e6-eb24-4f4f-b794-b11b496778e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.160042 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.160325 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178203e6-eb24-4f4f-b794-b11b496778e4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "178203e6-eb24-4f4f-b794-b11b496778e4" (UID: "178203e6-eb24-4f4f-b794-b11b496778e4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.160310 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178203e6-eb24-4f4f-b794-b11b496778e4-kube-api-access-jrt2s" (OuterVolumeSpecName: "kube-api-access-jrt2s") pod "178203e6-eb24-4f4f-b794-b11b496778e4" (UID: "178203e6-eb24-4f4f-b794-b11b496778e4"). InnerVolumeSpecName "kube-api-access-jrt2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.165478 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551350-vmx4v"] Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.241575 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl"] Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.245047 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.246834 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl"] Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.248412 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgvdq\" (UniqueName: \"kubernetes.io/projected/a5f670b4-f234-4c50-beb3-5ae4d0492d3f-kube-api-access-mgvdq\") pod \"auto-csr-approver-29551350-vmx4v\" (UID: \"a5f670b4-f234-4c50-beb3-5ae4d0492d3f\") " pod="openshift-infra/auto-csr-approver-29551350-vmx4v" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.248507 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/178203e6-eb24-4f4f-b794-b11b496778e4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.248527 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.248539 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrt2s\" (UniqueName: \"kubernetes.io/projected/178203e6-eb24-4f4f-b794-b11b496778e4-kube-api-access-jrt2s\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.248557 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178203e6-eb24-4f4f-b794-b11b496778e4-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.248936 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.249314 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.349744 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1069377-fcc6-4478-92cb-a93487954a97-config-volume\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.349797 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln8pp\" (UniqueName: \"kubernetes.io/projected/b1069377-fcc6-4478-92cb-a93487954a97-kube-api-access-ln8pp\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.349853 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1069377-fcc6-4478-92cb-a93487954a97-secret-volume\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.349954 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgvdq\" (UniqueName: \"kubernetes.io/projected/a5f670b4-f234-4c50-beb3-5ae4d0492d3f-kube-api-access-mgvdq\") pod \"auto-csr-approver-29551350-vmx4v\" (UID: \"a5f670b4-f234-4c50-beb3-5ae4d0492d3f\") " pod="openshift-infra/auto-csr-approver-29551350-vmx4v" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.370487 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgvdq\" (UniqueName: \"kubernetes.io/projected/a5f670b4-f234-4c50-beb3-5ae4d0492d3f-kube-api-access-mgvdq\") pod \"auto-csr-approver-29551350-vmx4v\" (UID: \"a5f670b4-f234-4c50-beb3-5ae4d0492d3f\") " pod="openshift-infra/auto-csr-approver-29551350-vmx4v" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.451892 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1069377-fcc6-4478-92cb-a93487954a97-config-volume\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.451965 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln8pp\" (UniqueName: \"kubernetes.io/projected/b1069377-fcc6-4478-92cb-a93487954a97-kube-api-access-ln8pp\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.452078 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1069377-fcc6-4478-92cb-a93487954a97-secret-volume\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.453438 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1069377-fcc6-4478-92cb-a93487954a97-config-volume\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.459197 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1069377-fcc6-4478-92cb-a93487954a97-secret-volume\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.471980 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln8pp\" (UniqueName: \"kubernetes.io/projected/b1069377-fcc6-4478-92cb-a93487954a97-kube-api-access-ln8pp\") pod \"collect-profiles-29551350-2zzrl\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.493423 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551350-vmx4v" Mar 09 18:30:00 crc kubenswrapper[4750]: I0309 18:30:00.574184 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:01 crc kubenswrapper[4750]: I0309 18:30:01.058101 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79" Mar 09 18:30:01 crc kubenswrapper[4750]: I0309 18:30:01.087738 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79"] Mar 09 18:30:01 crc kubenswrapper[4750]: I0309 18:30:01.094136 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dc67c9bf-vzb79"] Mar 09 18:30:01 crc kubenswrapper[4750]: I0309 18:30:01.385571 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="178203e6-eb24-4f4f-b794-b11b496778e4" path="/var/lib/kubelet/pods/178203e6-eb24-4f4f-b794-b11b496778e4/volumes" Mar 09 18:30:03 crc kubenswrapper[4750]: I0309 18:30:03.215700 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-svpt8" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.476116 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b874b4dd8-bhvkp"] Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.617401 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj"] Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.618236 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.621428 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.621983 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.622175 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.622297 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.622461 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.623006 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.630991 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj"] Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.715686 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-client-ca\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.715751 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-serving-cert\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.715845 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-config\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.715887 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvs88\" (UniqueName: \"kubernetes.io/projected/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-kube-api-access-xvs88\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.817544 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-config\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.817989 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvs88\" (UniqueName: \"kubernetes.io/projected/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-kube-api-access-xvs88\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.818130 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-client-ca\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.818237 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-serving-cert\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.819017 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-client-ca\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.819805 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-config\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.831818 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-serving-cert\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.837966 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvs88\" (UniqueName: \"kubernetes.io/projected/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-kube-api-access-xvs88\") pod \"route-controller-manager-59d9b4ccbf-tfjzj\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:04 crc kubenswrapper[4750]: I0309 18:30:04.936545 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:07 crc kubenswrapper[4750]: E0309 18:30:07.820537 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 09 18:30:07 crc kubenswrapper[4750]: E0309 18:30:07.821332 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dn6pd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-76nqh_openshift-marketplace(3c35f63c-cbc8-4d2b-b971-3e18a8222bc7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:30:07 crc kubenswrapper[4750]: E0309 18:30:07.822645 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-76nqh" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" Mar 09 18:30:07 crc kubenswrapper[4750]: E0309 18:30:07.870163 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 09 18:30:07 crc kubenswrapper[4750]: E0309 18:30:07.870347 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lw7mj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lxwd2_openshift-marketplace(8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:30:07 crc kubenswrapper[4750]: E0309 18:30:07.871540 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lxwd2" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.521124 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.526429 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.527187 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.535249 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.535259 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.593660 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d4870854-2a1f-497c-8460-96bb72c36360-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.594145 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4870854-2a1f-497c-8460-96bb72c36360-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.695420 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d4870854-2a1f-497c-8460-96bb72c36360-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.695483 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4870854-2a1f-497c-8460-96bb72c36360-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.696062 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d4870854-2a1f-497c-8460-96bb72c36360-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.713252 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.724212 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4870854-2a1f-497c-8460-96bb72c36360-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:09 crc kubenswrapper[4750]: I0309 18:30:09.859990 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:10 crc kubenswrapper[4750]: E0309 18:30:10.018207 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lxwd2" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" Mar 09 18:30:10 crc kubenswrapper[4750]: E0309 18:30:10.018293 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-76nqh" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" Mar 09 18:30:10 crc kubenswrapper[4750]: E0309 18:30:10.099908 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 09 18:30:10 crc kubenswrapper[4750]: E0309 18:30:10.100130 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c66jx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dk5bm_openshift-marketplace(7c78295a-888f-4690-b970-0e1557de4ad5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:30:10 crc kubenswrapper[4750]: E0309 18:30:10.101376 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dk5bm" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.643060 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dk5bm" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.753084 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.753781 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-878gv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rj4zw_openshift-marketplace(41608683-d508-4374-a570-0e93ca339ffb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.763468 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rj4zw" podUID="41608683-d508-4374-a570-0e93ca339ffb" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.875242 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.877186 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-btsjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zvzmg_openshift-marketplace(ab00ccba-a824-42a2-a9fc-43139bcf9a0e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.879763 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zvzmg" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.966875 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.967030 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9k5lr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-v879l_openshift-marketplace(b880ae0c-0f6f-487b-8927-77437bb8a638): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.975598 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.975828 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2ns2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-86fld_openshift-marketplace(3418a0da-42b0-4ac4-8957-87ed04209742): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.977824 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-v879l" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" Mar 09 18:30:11 crc kubenswrapper[4750]: E0309 18:30:11.977939 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-86fld" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.146038 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk4f9" event={"ID":"2de1257e-0c9d-486f-92da-2e0525941f7e","Type":"ContainerStarted","Data":"425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b"} Mar 09 18:30:12 crc kubenswrapper[4750]: E0309 18:30:12.150328 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rj4zw" podUID="41608683-d508-4374-a570-0e93ca339ffb" Mar 09 18:30:12 crc kubenswrapper[4750]: E0309 18:30:12.150718 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zvzmg" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" Mar 09 18:30:12 crc kubenswrapper[4750]: E0309 18:30:12.151586 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86fld" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" Mar 09 18:30:12 crc kubenswrapper[4750]: E0309 18:30:12.152112 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-v879l" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.300682 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b874b4dd8-bhvkp"] Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.340803 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl"] Mar 09 18:30:12 crc kubenswrapper[4750]: W0309 18:30:12.360532 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44a0b9c1_cf55_47d6_abf1_d0d6529fb3a6.slice/crio-3966b2bcbf6e2ae9eaddadbf8c682d5bc162b810fc62cdba7c0b662fa921eddd WatchSource:0}: Error finding container 3966b2bcbf6e2ae9eaddadbf8c682d5bc162b810fc62cdba7c0b662fa921eddd: Status 404 returned error can't find the container with id 3966b2bcbf6e2ae9eaddadbf8c682d5bc162b810fc62cdba7c0b662fa921eddd Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.364022 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551350-vmx4v"] Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.380247 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj"] Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.386478 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 09 18:30:12 crc kubenswrapper[4750]: W0309 18:30:12.397029 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd4870854_2a1f_497c_8460_96bb72c36360.slice/crio-ad5ea324f40a3868bde7fbdba6a35677b52f04728b8ed521df2ab87fe614b5ab WatchSource:0}: Error finding container ad5ea324f40a3868bde7fbdba6a35677b52f04728b8ed521df2ab87fe614b5ab: Status 404 returned error can't find the container with id ad5ea324f40a3868bde7fbdba6a35677b52f04728b8ed521df2ab87fe614b5ab Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.882768 4750 csr.go:261] certificate signing request csr-qfdkh is approved, waiting to be issued Mar 09 18:30:12 crc kubenswrapper[4750]: I0309 18:30:12.892223 4750 csr.go:257] certificate signing request csr-qfdkh is issued Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.156958 4750 generic.go:334] "Generic (PLEG): container finished" podID="a08fcb79-42d8-47c2-940e-92c4ff4f2f61" containerID="06cecdc078351cbf290c74ea9df45ed62c823161e4a770caa783c20d20debcb0" exitCode=0 Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.157161 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" event={"ID":"a08fcb79-42d8-47c2-940e-92c4ff4f2f61","Type":"ContainerDied","Data":"06cecdc078351cbf290c74ea9df45ed62c823161e4a770caa783c20d20debcb0"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.160270 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" event={"ID":"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7","Type":"ContainerStarted","Data":"a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.160349 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" event={"ID":"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7","Type":"ContainerStarted","Data":"4b1c000d8b59903d443144362681c6fb8256986a261059f7320db74e70717a2e"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.160469 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.162811 4750 generic.go:334] "Generic (PLEG): container finished" podID="b1069377-fcc6-4478-92cb-a93487954a97" containerID="1466913896df81ea73a2eee20a325bdc8813bed7a46c93fee0925d84030bd21a" exitCode=0 Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.162914 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" event={"ID":"b1069377-fcc6-4478-92cb-a93487954a97","Type":"ContainerDied","Data":"1466913896df81ea73a2eee20a325bdc8813bed7a46c93fee0925d84030bd21a"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.162946 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" event={"ID":"b1069377-fcc6-4478-92cb-a93487954a97","Type":"ContainerStarted","Data":"596a5a958664575423d2804786937c380b35ccd860e851b2e461e2fdcb8a8069"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.165800 4750 generic.go:334] "Generic (PLEG): container finished" podID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerID="425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b" exitCode=0 Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.165893 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk4f9" event={"ID":"2de1257e-0c9d-486f-92da-2e0525941f7e","Type":"ContainerDied","Data":"425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.167961 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d4870854-2a1f-497c-8460-96bb72c36360","Type":"ContainerStarted","Data":"8bcfaacb7aecca0461f2827dd263c81311db36f0326cd1784a2c3a423ec9217c"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.168013 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d4870854-2a1f-497c-8460-96bb72c36360","Type":"ContainerStarted","Data":"ad5ea324f40a3868bde7fbdba6a35677b52f04728b8ed521df2ab87fe614b5ab"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.169613 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" event={"ID":"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6","Type":"ContainerStarted","Data":"4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.169673 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" event={"ID":"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6","Type":"ContainerStarted","Data":"3966b2bcbf6e2ae9eaddadbf8c682d5bc162b810fc62cdba7c0b662fa921eddd"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.169759 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" podUID="44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" containerName="controller-manager" containerID="cri-o://4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5" gracePeriod=30 Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.169870 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.172882 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551350-vmx4v" event={"ID":"a5f670b4-f234-4c50-beb3-5ae4d0492d3f","Type":"ContainerStarted","Data":"6acf26725b8de70537720f809112c0e89e09cfb8a98c9e89883655d888580ea6"} Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.179796 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.187319 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.206861 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" podStartSLOduration=29.206836827 podStartE2EDuration="29.206836827s" podCreationTimestamp="2026-03-09 18:29:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:13.199744014 +0000 UTC m=+294.542216432" watchObservedRunningTime="2026-03-09 18:30:13.206836827 +0000 UTC m=+294.549309225" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.246728 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" podStartSLOduration=9.246702607 podStartE2EDuration="9.246702607s" podCreationTimestamp="2026-03-09 18:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:13.243570651 +0000 UTC m=+294.586043049" watchObservedRunningTime="2026-03-09 18:30:13.246702607 +0000 UTC m=+294.589175005" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.258946 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=4.258922191 podStartE2EDuration="4.258922191s" podCreationTimestamp="2026-03-09 18:30:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:13.258339274 +0000 UTC m=+294.600811692" watchObservedRunningTime="2026-03-09 18:30:13.258922191 +0000 UTC m=+294.601394589" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.653581 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.690144 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7fc6977f7f-rz828"] Mar 09 18:30:13 crc kubenswrapper[4750]: E0309 18:30:13.690542 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" containerName="controller-manager" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.690564 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" containerName="controller-manager" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.690709 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" containerName="controller-manager" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.691409 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.695716 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7fc6977f7f-rz828"] Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.781664 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dswp\" (UniqueName: \"kubernetes.io/projected/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-kube-api-access-8dswp\") pod \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782220 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-serving-cert\") pod \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782261 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-proxy-ca-bundles\") pod \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782353 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-config\") pod \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782399 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-client-ca\") pod \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\" (UID: \"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6\") " Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782597 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skdlq\" (UniqueName: \"kubernetes.io/projected/eb6df7f3-70c7-4913-b305-be30b8f2a48b-kube-api-access-skdlq\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782674 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-config\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782746 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb6df7f3-70c7-4913-b305-be30b8f2a48b-serving-cert\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782790 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-proxy-ca-bundles\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.782816 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-client-ca\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.783255 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" (UID: "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.783952 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-config" (OuterVolumeSpecName: "config") pod "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" (UID: "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.783992 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" (UID: "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.790286 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-kube-api-access-8dswp" (OuterVolumeSpecName: "kube-api-access-8dswp") pod "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" (UID: "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6"). InnerVolumeSpecName "kube-api-access-8dswp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.791285 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" (UID: "44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884225 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb6df7f3-70c7-4913-b305-be30b8f2a48b-serving-cert\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884321 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-proxy-ca-bundles\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884350 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-client-ca\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884382 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skdlq\" (UniqueName: \"kubernetes.io/projected/eb6df7f3-70c7-4913-b305-be30b8f2a48b-kube-api-access-skdlq\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884411 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-config\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884457 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dswp\" (UniqueName: \"kubernetes.io/projected/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-kube-api-access-8dswp\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884470 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884481 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884492 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.884502 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.886320 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-client-ca\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.886681 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-proxy-ca-bundles\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.887160 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-config\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.888937 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb6df7f3-70c7-4913-b305-be30b8f2a48b-serving-cert\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.894091 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-30 18:10:51.794801157 +0000 UTC Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.894136 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6383h40m37.900669964s for next certificate rotation Mar 09 18:30:13 crc kubenswrapper[4750]: I0309 18:30:13.916747 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skdlq\" (UniqueName: \"kubernetes.io/projected/eb6df7f3-70c7-4913-b305-be30b8f2a48b-kube-api-access-skdlq\") pod \"controller-manager-7fc6977f7f-rz828\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.023500 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.185083 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk4f9" event={"ID":"2de1257e-0c9d-486f-92da-2e0525941f7e","Type":"ContainerStarted","Data":"552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a"} Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.187137 4750 generic.go:334] "Generic (PLEG): container finished" podID="d4870854-2a1f-497c-8460-96bb72c36360" containerID="8bcfaacb7aecca0461f2827dd263c81311db36f0326cd1784a2c3a423ec9217c" exitCode=0 Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.187256 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d4870854-2a1f-497c-8460-96bb72c36360","Type":"ContainerDied","Data":"8bcfaacb7aecca0461f2827dd263c81311db36f0326cd1784a2c3a423ec9217c"} Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.189765 4750 generic.go:334] "Generic (PLEG): container finished" podID="44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" containerID="4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5" exitCode=0 Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.189888 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.190600 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" event={"ID":"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6","Type":"ContainerDied","Data":"4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5"} Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.190700 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b874b4dd8-bhvkp" event={"ID":"44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6","Type":"ContainerDied","Data":"3966b2bcbf6e2ae9eaddadbf8c682d5bc162b810fc62cdba7c0b662fa921eddd"} Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.190727 4750 scope.go:117] "RemoveContainer" containerID="4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.202602 4750 generic.go:334] "Generic (PLEG): container finished" podID="a5f670b4-f234-4c50-beb3-5ae4d0492d3f" containerID="4251eb29c736506ce6e889790879763b504a1939bdae6db77b810fae289f67a4" exitCode=0 Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.202713 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551350-vmx4v" event={"ID":"a5f670b4-f234-4c50-beb3-5ae4d0492d3f","Type":"ContainerDied","Data":"4251eb29c736506ce6e889790879763b504a1939bdae6db77b810fae289f67a4"} Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.216383 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jk4f9" podStartSLOduration=4.065968842 podStartE2EDuration="42.216358312s" podCreationTimestamp="2026-03-09 18:29:32 +0000 UTC" firstStartedPulling="2026-03-09 18:29:35.590883555 +0000 UTC m=+256.933355953" lastFinishedPulling="2026-03-09 18:30:13.741273025 +0000 UTC m=+295.083745423" observedRunningTime="2026-03-09 18:30:14.216333031 +0000 UTC m=+295.558805429" watchObservedRunningTime="2026-03-09 18:30:14.216358312 +0000 UTC m=+295.558830710" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.246136 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7fc6977f7f-rz828"] Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.254126 4750 scope.go:117] "RemoveContainer" containerID="4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5" Mar 09 18:30:14 crc kubenswrapper[4750]: E0309 18:30:14.256590 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5\": container with ID starting with 4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5 not found: ID does not exist" containerID="4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.256711 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5"} err="failed to get container status \"4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5\": rpc error: code = NotFound desc = could not find container \"4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5\": container with ID starting with 4f8e1a527461113d83647b73cdf01b33319f84337980d261ac2469e8aa521fe5 not found: ID does not exist" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.273963 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b874b4dd8-bhvkp"] Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.276475 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b874b4dd8-bhvkp"] Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.450365 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.517161 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 09 18:30:14 crc kubenswrapper[4750]: E0309 18:30:14.517936 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1069377-fcc6-4478-92cb-a93487954a97" containerName="collect-profiles" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.517956 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1069377-fcc6-4478-92cb-a93487954a97" containerName="collect-profiles" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.518196 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1069377-fcc6-4478-92cb-a93487954a97" containerName="collect-profiles" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.518784 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.522259 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.525886 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.593682 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1069377-fcc6-4478-92cb-a93487954a97-secret-volume\") pod \"b1069377-fcc6-4478-92cb-a93487954a97\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.593862 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln8pp\" (UniqueName: \"kubernetes.io/projected/b1069377-fcc6-4478-92cb-a93487954a97-kube-api-access-ln8pp\") pod \"b1069377-fcc6-4478-92cb-a93487954a97\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.593919 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1069377-fcc6-4478-92cb-a93487954a97-config-volume\") pod \"b1069377-fcc6-4478-92cb-a93487954a97\" (UID: \"b1069377-fcc6-4478-92cb-a93487954a97\") " Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.594037 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk6ng\" (UniqueName: \"kubernetes.io/projected/a08fcb79-42d8-47c2-940e-92c4ff4f2f61-kube-api-access-fk6ng\") pod \"a08fcb79-42d8-47c2-940e-92c4ff4f2f61\" (UID: \"a08fcb79-42d8-47c2-940e-92c4ff4f2f61\") " Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.594225 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kubelet-dir\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.594280 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kube-api-access\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.594390 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-var-lock\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.595546 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1069377-fcc6-4478-92cb-a93487954a97-config-volume" (OuterVolumeSpecName: "config-volume") pod "b1069377-fcc6-4478-92cb-a93487954a97" (UID: "b1069377-fcc6-4478-92cb-a93487954a97"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.599825 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1069377-fcc6-4478-92cb-a93487954a97-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b1069377-fcc6-4478-92cb-a93487954a97" (UID: "b1069377-fcc6-4478-92cb-a93487954a97"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.600457 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a08fcb79-42d8-47c2-940e-92c4ff4f2f61-kube-api-access-fk6ng" (OuterVolumeSpecName: "kube-api-access-fk6ng") pod "a08fcb79-42d8-47c2-940e-92c4ff4f2f61" (UID: "a08fcb79-42d8-47c2-940e-92c4ff4f2f61"). InnerVolumeSpecName "kube-api-access-fk6ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.600582 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1069377-fcc6-4478-92cb-a93487954a97-kube-api-access-ln8pp" (OuterVolumeSpecName: "kube-api-access-ln8pp") pod "b1069377-fcc6-4478-92cb-a93487954a97" (UID: "b1069377-fcc6-4478-92cb-a93487954a97"). InnerVolumeSpecName "kube-api-access-ln8pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.695807 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-var-lock\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.695894 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kubelet-dir\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.695922 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kube-api-access\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.695976 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk6ng\" (UniqueName: \"kubernetes.io/projected/a08fcb79-42d8-47c2-940e-92c4ff4f2f61-kube-api-access-fk6ng\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.695988 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1069377-fcc6-4478-92cb-a93487954a97-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.695999 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln8pp\" (UniqueName: \"kubernetes.io/projected/b1069377-fcc6-4478-92cb-a93487954a97-kube-api-access-ln8pp\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.696008 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1069377-fcc6-4478-92cb-a93487954a97-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.696358 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-var-lock\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.696391 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kubelet-dir\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.736759 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kube-api-access\") pod \"installer-9-crc\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.833521 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.894819 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-08 17:14:01.562472122 +0000 UTC Mar 09 18:30:14 crc kubenswrapper[4750]: I0309 18:30:14.894870 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7318h43m46.667604995s for next certificate rotation Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.058081 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 09 18:30:15 crc kubenswrapper[4750]: W0309 18:30:15.065817 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8e321a00_cd51_4fc5_b20f_8a6c5d7f3243.slice/crio-e0edc50f0132806189b063abe684e4b58ff151e63c4ec71a441b986483858788 WatchSource:0}: Error finding container e0edc50f0132806189b063abe684e4b58ff151e63c4ec71a441b986483858788: Status 404 returned error can't find the container with id e0edc50f0132806189b063abe684e4b58ff151e63c4ec71a441b986483858788 Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.213560 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243","Type":"ContainerStarted","Data":"e0edc50f0132806189b063abe684e4b58ff151e63c4ec71a441b986483858788"} Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.216653 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" event={"ID":"eb6df7f3-70c7-4913-b305-be30b8f2a48b","Type":"ContainerStarted","Data":"f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0"} Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.216717 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" event={"ID":"eb6df7f3-70c7-4913-b305-be30b8f2a48b","Type":"ContainerStarted","Data":"e4654d620ecc15d15512cd6ed30745d8d705fcdd82604089a4d948b6570f1b91"} Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.217007 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.220300 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" event={"ID":"a08fcb79-42d8-47c2-940e-92c4ff4f2f61","Type":"ContainerDied","Data":"a447fc03882759db56727d5053167aa55d24d12a915c5d97fc7a09caddb70110"} Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.220343 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a447fc03882759db56727d5053167aa55d24d12a915c5d97fc7a09caddb70110" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.220536 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551348-g9vjb" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.222577 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" event={"ID":"b1069377-fcc6-4478-92cb-a93487954a97","Type":"ContainerDied","Data":"596a5a958664575423d2804786937c380b35ccd860e851b2e461e2fdcb8a8069"} Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.222616 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="596a5a958664575423d2804786937c380b35ccd860e851b2e461e2fdcb8a8069" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.222665 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.249947 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.274167 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" podStartSLOduration=11.274144834 podStartE2EDuration="11.274144834s" podCreationTimestamp="2026-03-09 18:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:15.24909631 +0000 UTC m=+296.591568708" watchObservedRunningTime="2026-03-09 18:30:15.274144834 +0000 UTC m=+296.616617232" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.399190 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6" path="/var/lib/kubelet/pods/44a0b9c1-cf55-47d6-abf1-d0d6529fb3a6/volumes" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.505034 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.536018 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551350-vmx4v" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.611305 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgvdq\" (UniqueName: \"kubernetes.io/projected/a5f670b4-f234-4c50-beb3-5ae4d0492d3f-kube-api-access-mgvdq\") pod \"a5f670b4-f234-4c50-beb3-5ae4d0492d3f\" (UID: \"a5f670b4-f234-4c50-beb3-5ae4d0492d3f\") " Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.611393 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4870854-2a1f-497c-8460-96bb72c36360-kube-api-access\") pod \"d4870854-2a1f-497c-8460-96bb72c36360\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.611440 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d4870854-2a1f-497c-8460-96bb72c36360-kubelet-dir\") pod \"d4870854-2a1f-497c-8460-96bb72c36360\" (UID: \"d4870854-2a1f-497c-8460-96bb72c36360\") " Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.611587 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4870854-2a1f-497c-8460-96bb72c36360-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d4870854-2a1f-497c-8460-96bb72c36360" (UID: "d4870854-2a1f-497c-8460-96bb72c36360"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.611761 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d4870854-2a1f-497c-8460-96bb72c36360-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.617164 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4870854-2a1f-497c-8460-96bb72c36360-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d4870854-2a1f-497c-8460-96bb72c36360" (UID: "d4870854-2a1f-497c-8460-96bb72c36360"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.619505 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f670b4-f234-4c50-beb3-5ae4d0492d3f-kube-api-access-mgvdq" (OuterVolumeSpecName: "kube-api-access-mgvdq") pod "a5f670b4-f234-4c50-beb3-5ae4d0492d3f" (UID: "a5f670b4-f234-4c50-beb3-5ae4d0492d3f"). InnerVolumeSpecName "kube-api-access-mgvdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.712786 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgvdq\" (UniqueName: \"kubernetes.io/projected/a5f670b4-f234-4c50-beb3-5ae4d0492d3f-kube-api-access-mgvdq\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:15 crc kubenswrapper[4750]: I0309 18:30:15.712827 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4870854-2a1f-497c-8460-96bb72c36360-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.230647 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d4870854-2a1f-497c-8460-96bb72c36360","Type":"ContainerDied","Data":"ad5ea324f40a3868bde7fbdba6a35677b52f04728b8ed521df2ab87fe614b5ab"} Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.230706 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad5ea324f40a3868bde7fbdba6a35677b52f04728b8ed521df2ab87fe614b5ab" Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.230664 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.232126 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243","Type":"ContainerStarted","Data":"48f6a8232e40d5d679843d5bc884eae21d8c36abeb7755e85aa03dfb1ede06cf"} Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.233862 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551350-vmx4v" event={"ID":"a5f670b4-f234-4c50-beb3-5ae4d0492d3f","Type":"ContainerDied","Data":"6acf26725b8de70537720f809112c0e89e09cfb8a98c9e89883655d888580ea6"} Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.233891 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551350-vmx4v" Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.233912 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6acf26725b8de70537720f809112c0e89e09cfb8a98c9e89883655d888580ea6" Mar 09 18:30:16 crc kubenswrapper[4750]: I0309 18:30:16.270544 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.27051992 podStartE2EDuration="2.27051992s" podCreationTimestamp="2026-03-09 18:30:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:16.265011299 +0000 UTC m=+297.607483697" watchObservedRunningTime="2026-03-09 18:30:16.27051992 +0000 UTC m=+297.612992318" Mar 09 18:30:21 crc kubenswrapper[4750]: I0309 18:30:21.743563 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:30:21 crc kubenswrapper[4750]: I0309 18:30:21.744328 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:30:21 crc kubenswrapper[4750]: I0309 18:30:21.744444 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:30:21 crc kubenswrapper[4750]: I0309 18:30:21.745382 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:30:21 crc kubenswrapper[4750]: I0309 18:30:21.745449 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26" gracePeriod=600 Mar 09 18:30:22 crc kubenswrapper[4750]: I0309 18:30:22.271845 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26" exitCode=0 Mar 09 18:30:22 crc kubenswrapper[4750]: I0309 18:30:22.271963 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26"} Mar 09 18:30:22 crc kubenswrapper[4750]: I0309 18:30:22.272316 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"0818ed3c5ca0c09708e66182b106417995a2e2cc5b7ae8c6dec5d1157e85abf4"} Mar 09 18:30:23 crc kubenswrapper[4750]: I0309 18:30:23.143379 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:30:23 crc kubenswrapper[4750]: I0309 18:30:23.143915 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:30:23 crc kubenswrapper[4750]: I0309 18:30:23.477472 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:30:23 crc kubenswrapper[4750]: I0309 18:30:23.526994 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.287967 4750 generic.go:334] "Generic (PLEG): container finished" podID="3418a0da-42b0-4ac4-8957-87ed04209742" containerID="2b4f956de617c1de2bb545878d101e514a90c15330b64f5836c7a68b098f0d42" exitCode=0 Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.288055 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86fld" event={"ID":"3418a0da-42b0-4ac4-8957-87ed04209742","Type":"ContainerDied","Data":"2b4f956de617c1de2bb545878d101e514a90c15330b64f5836c7a68b098f0d42"} Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.294752 4750 generic.go:334] "Generic (PLEG): container finished" podID="7c78295a-888f-4690-b970-0e1557de4ad5" containerID="80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6" exitCode=0 Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.294814 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk5bm" event={"ID":"7c78295a-888f-4690-b970-0e1557de4ad5","Type":"ContainerDied","Data":"80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6"} Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.466322 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7fc6977f7f-rz828"] Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.466570 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" podUID="eb6df7f3-70c7-4913-b305-be30b8f2a48b" containerName="controller-manager" containerID="cri-o://f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0" gracePeriod=30 Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.483161 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj"] Mar 09 18:30:24 crc kubenswrapper[4750]: I0309 18:30:24.483394 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" podUID="abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" containerName="route-controller-manager" containerID="cri-o://a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259" gracePeriod=30 Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.053274 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.062846 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.151920 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb6df7f3-70c7-4913-b305-be30b8f2a48b-serving-cert\") pod \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.151994 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-config\") pod \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.152030 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-proxy-ca-bundles\") pod \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.152089 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-config\") pod \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.152147 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-client-ca\") pod \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.152201 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvs88\" (UniqueName: \"kubernetes.io/projected/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-kube-api-access-xvs88\") pod \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.152242 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-serving-cert\") pod \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\" (UID: \"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.152283 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-client-ca\") pod \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.152328 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skdlq\" (UniqueName: \"kubernetes.io/projected/eb6df7f3-70c7-4913-b305-be30b8f2a48b-kube-api-access-skdlq\") pod \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\" (UID: \"eb6df7f3-70c7-4913-b305-be30b8f2a48b\") " Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.153095 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "eb6df7f3-70c7-4913-b305-be30b8f2a48b" (UID: "eb6df7f3-70c7-4913-b305-be30b8f2a48b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.153285 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-client-ca" (OuterVolumeSpecName: "client-ca") pod "eb6df7f3-70c7-4913-b305-be30b8f2a48b" (UID: "eb6df7f3-70c7-4913-b305-be30b8f2a48b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.153327 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" (UID: "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.153361 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-config" (OuterVolumeSpecName: "config") pod "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" (UID: "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.154749 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-config" (OuterVolumeSpecName: "config") pod "eb6df7f3-70c7-4913-b305-be30b8f2a48b" (UID: "eb6df7f3-70c7-4913-b305-be30b8f2a48b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.159375 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6df7f3-70c7-4913-b305-be30b8f2a48b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "eb6df7f3-70c7-4913-b305-be30b8f2a48b" (UID: "eb6df7f3-70c7-4913-b305-be30b8f2a48b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.159820 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb6df7f3-70c7-4913-b305-be30b8f2a48b-kube-api-access-skdlq" (OuterVolumeSpecName: "kube-api-access-skdlq") pod "eb6df7f3-70c7-4913-b305-be30b8f2a48b" (UID: "eb6df7f3-70c7-4913-b305-be30b8f2a48b"). InnerVolumeSpecName "kube-api-access-skdlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.160111 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" (UID: "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.160287 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-kube-api-access-xvs88" (OuterVolumeSpecName: "kube-api-access-xvs88") pod "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" (UID: "abb06a6d-cec1-41f8-9c38-bffd1eadd6b7"). InnerVolumeSpecName "kube-api-access-xvs88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.253976 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254036 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvs88\" (UniqueName: \"kubernetes.io/projected/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-kube-api-access-xvs88\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254050 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254062 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254074 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skdlq\" (UniqueName: \"kubernetes.io/projected/eb6df7f3-70c7-4913-b305-be30b8f2a48b-kube-api-access-skdlq\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254083 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb6df7f3-70c7-4913-b305-be30b8f2a48b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254106 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254115 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.254126 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6df7f3-70c7-4913-b305-be30b8f2a48b-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.303031 4750 generic.go:334] "Generic (PLEG): container finished" podID="eb6df7f3-70c7-4913-b305-be30b8f2a48b" containerID="f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0" exitCode=0 Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.303103 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" event={"ID":"eb6df7f3-70c7-4913-b305-be30b8f2a48b","Type":"ContainerDied","Data":"f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.303136 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" event={"ID":"eb6df7f3-70c7-4913-b305-be30b8f2a48b","Type":"ContainerDied","Data":"e4654d620ecc15d15512cd6ed30745d8d705fcdd82604089a4d948b6570f1b91"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.303158 4750 scope.go:117] "RemoveContainer" containerID="f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.303275 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fc6977f7f-rz828" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.312166 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk5bm" event={"ID":"7c78295a-888f-4690-b970-0e1557de4ad5","Type":"ContainerStarted","Data":"2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.315191 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxwd2" event={"ID":"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082","Type":"ContainerStarted","Data":"8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.317240 4750 generic.go:334] "Generic (PLEG): container finished" podID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerID="0afceb5bb80b11f325cf4af12963ca64b215c4d515f87e58ff50400fbb789897" exitCode=0 Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.317311 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76nqh" event={"ID":"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7","Type":"ContainerDied","Data":"0afceb5bb80b11f325cf4af12963ca64b215c4d515f87e58ff50400fbb789897"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.319215 4750 generic.go:334] "Generic (PLEG): container finished" podID="abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" containerID="a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259" exitCode=0 Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.319276 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.319313 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" event={"ID":"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7","Type":"ContainerDied","Data":"a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.319348 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" event={"ID":"abb06a6d-cec1-41f8-9c38-bffd1eadd6b7","Type":"ContainerDied","Data":"4b1c000d8b59903d443144362681c6fb8256986a261059f7320db74e70717a2e"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.321780 4750 scope.go:117] "RemoveContainer" containerID="f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0" Mar 09 18:30:25 crc kubenswrapper[4750]: E0309 18:30:25.322305 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0\": container with ID starting with f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0 not found: ID does not exist" containerID="f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.322355 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0"} err="failed to get container status \"f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0\": rpc error: code = NotFound desc = could not find container \"f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0\": container with ID starting with f3470ca51ca97b15da9ee95e31eebf1081fddf2d6dc8113cc44f2f96d757deb0 not found: ID does not exist" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.322393 4750 scope.go:117] "RemoveContainer" containerID="a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.323279 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86fld" event={"ID":"3418a0da-42b0-4ac4-8957-87ed04209742","Type":"ContainerStarted","Data":"2923e088f5d0b051431ec290da04e6ee201bdfd710e3b95ebc8b3a151d27bd97"} Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.344694 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dk5bm" podStartSLOduration=2.670415852 podStartE2EDuration="54.34459974s" podCreationTimestamp="2026-03-09 18:29:31 +0000 UTC" firstStartedPulling="2026-03-09 18:29:33.174095365 +0000 UTC m=+254.516567763" lastFinishedPulling="2026-03-09 18:30:24.848279253 +0000 UTC m=+306.190751651" observedRunningTime="2026-03-09 18:30:25.339297405 +0000 UTC m=+306.681769803" watchObservedRunningTime="2026-03-09 18:30:25.34459974 +0000 UTC m=+306.687072158" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.358556 4750 scope.go:117] "RemoveContainer" containerID="a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259" Mar 09 18:30:25 crc kubenswrapper[4750]: E0309 18:30:25.359144 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259\": container with ID starting with a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259 not found: ID does not exist" containerID="a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.359175 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259"} err="failed to get container status \"a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259\": rpc error: code = NotFound desc = could not find container \"a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259\": container with ID starting with a0b7dad0e6046e146aa16986a703504271fb5f0b6ced0c6cd68ce6bb00110259 not found: ID does not exist" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.367181 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-86fld" podStartSLOduration=2.656944672 podStartE2EDuration="54.367168556s" podCreationTimestamp="2026-03-09 18:29:31 +0000 UTC" firstStartedPulling="2026-03-09 18:29:33.070730199 +0000 UTC m=+254.413202597" lastFinishedPulling="2026-03-09 18:30:24.780954083 +0000 UTC m=+306.123426481" observedRunningTime="2026-03-09 18:30:25.361698117 +0000 UTC m=+306.704170515" watchObservedRunningTime="2026-03-09 18:30:25.367168556 +0000 UTC m=+306.709640954" Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.453174 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7fc6977f7f-rz828"] Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.456078 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7fc6977f7f-rz828"] Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.491174 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj"] Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.496460 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj"] Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.937831 4750 patch_prober.go:28] interesting pod/route-controller-manager-59d9b4ccbf-tfjzj container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 18:30:25 crc kubenswrapper[4750]: I0309 18:30:25.938327 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-59d9b4ccbf-tfjzj" podUID="abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.332083 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvzmg" event={"ID":"ab00ccba-a824-42a2-a9fc-43139bcf9a0e","Type":"ContainerStarted","Data":"97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426"} Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.334043 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76nqh" event={"ID":"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7","Type":"ContainerStarted","Data":"2e7fdef072a9836002dbfdbce3591c960bf426ed1c793fe197e966459a1f1c6a"} Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.338876 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rj4zw" event={"ID":"41608683-d508-4374-a570-0e93ca339ffb","Type":"ContainerStarted","Data":"8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585"} Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.340461 4750 generic.go:334] "Generic (PLEG): container finished" podID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerID="8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c" exitCode=0 Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.340494 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxwd2" event={"ID":"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082","Type":"ContainerDied","Data":"8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c"} Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.379102 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-859c744c7b-skzpn"] Mar 09 18:30:26 crc kubenswrapper[4750]: E0309 18:30:26.379814 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb6df7f3-70c7-4913-b305-be30b8f2a48b" containerName="controller-manager" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.379918 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb6df7f3-70c7-4913-b305-be30b8f2a48b" containerName="controller-manager" Mar 09 18:30:26 crc kubenswrapper[4750]: E0309 18:30:26.379992 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f670b4-f234-4c50-beb3-5ae4d0492d3f" containerName="oc" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.380073 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f670b4-f234-4c50-beb3-5ae4d0492d3f" containerName="oc" Mar 09 18:30:26 crc kubenswrapper[4750]: E0309 18:30:26.380159 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a08fcb79-42d8-47c2-940e-92c4ff4f2f61" containerName="oc" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.380232 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a08fcb79-42d8-47c2-940e-92c4ff4f2f61" containerName="oc" Mar 09 18:30:26 crc kubenswrapper[4750]: E0309 18:30:26.380312 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4870854-2a1f-497c-8460-96bb72c36360" containerName="pruner" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.380381 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4870854-2a1f-497c-8460-96bb72c36360" containerName="pruner" Mar 09 18:30:26 crc kubenswrapper[4750]: E0309 18:30:26.380456 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" containerName="route-controller-manager" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.380528 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" containerName="route-controller-manager" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.380748 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f670b4-f234-4c50-beb3-5ae4d0492d3f" containerName="oc" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.380825 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a08fcb79-42d8-47c2-940e-92c4ff4f2f61" containerName="oc" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.380883 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb6df7f3-70c7-4913-b305-be30b8f2a48b" containerName="controller-manager" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.381022 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4870854-2a1f-497c-8460-96bb72c36360" containerName="pruner" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.381089 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" containerName="route-controller-manager" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.381700 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.382249 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k"] Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.383212 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.397330 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.397775 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398238 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398409 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398457 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398587 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398658 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398693 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398824 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.398995 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.399374 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.401264 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.405692 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k"] Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.414365 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-859c744c7b-skzpn"] Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.427046 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-76nqh" podStartSLOduration=3.336647754 podStartE2EDuration="52.427022356s" podCreationTimestamp="2026-03-09 18:29:34 +0000 UTC" firstStartedPulling="2026-03-09 18:29:36.674283129 +0000 UTC m=+258.016755527" lastFinishedPulling="2026-03-09 18:30:25.764657731 +0000 UTC m=+307.107130129" observedRunningTime="2026-03-09 18:30:26.424840726 +0000 UTC m=+307.767313124" watchObservedRunningTime="2026-03-09 18:30:26.427022356 +0000 UTC m=+307.769494744" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.428437 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473364 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-client-ca\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473443 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5872c77f-c7ee-4c4c-87c1-98fccae853c8-serving-cert\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473471 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-client-ca\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473501 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-config\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473544 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb82741c-91a2-49b9-b8f4-dd91167ec107-serving-cert\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473587 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-proxy-ca-bundles\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473616 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqj8z\" (UniqueName: \"kubernetes.io/projected/eb82741c-91a2-49b9-b8f4-dd91167ec107-kube-api-access-dqj8z\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473660 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fjq5\" (UniqueName: \"kubernetes.io/projected/5872c77f-c7ee-4c4c-87c1-98fccae853c8-kube-api-access-6fjq5\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.473688 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-config\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.575239 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqj8z\" (UniqueName: \"kubernetes.io/projected/eb82741c-91a2-49b9-b8f4-dd91167ec107-kube-api-access-dqj8z\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.575760 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fjq5\" (UniqueName: \"kubernetes.io/projected/5872c77f-c7ee-4c4c-87c1-98fccae853c8-kube-api-access-6fjq5\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.575792 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-config\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.577433 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-config\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.577614 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-client-ca\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.578428 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-client-ca\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.578611 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5872c77f-c7ee-4c4c-87c1-98fccae853c8-serving-cert\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.579748 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-client-ca\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.579818 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-config\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.579889 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb82741c-91a2-49b9-b8f4-dd91167ec107-serving-cert\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.579966 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-proxy-ca-bundles\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.580666 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-client-ca\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.581146 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-proxy-ca-bundles\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.581783 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-config\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.590382 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5872c77f-c7ee-4c4c-87c1-98fccae853c8-serving-cert\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.599253 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fjq5\" (UniqueName: \"kubernetes.io/projected/5872c77f-c7ee-4c4c-87c1-98fccae853c8-kube-api-access-6fjq5\") pod \"controller-manager-859c744c7b-skzpn\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.599740 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb82741c-91a2-49b9-b8f4-dd91167ec107-serving-cert\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.605384 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqj8z\" (UniqueName: \"kubernetes.io/projected/eb82741c-91a2-49b9-b8f4-dd91167ec107-kube-api-access-dqj8z\") pod \"route-controller-manager-7649f4cc88-6dg2k\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.697417 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.708914 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:26 crc kubenswrapper[4750]: I0309 18:30:26.984908 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-859c744c7b-skzpn"] Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.258684 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k"] Mar 09 18:30:27 crc kubenswrapper[4750]: W0309 18:30:27.263125 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb82741c_91a2_49b9_b8f4_dd91167ec107.slice/crio-1d383dc0b0d21b7ed190a52d6d55a9c12d04997a6febac24c97d76b6f496012d WatchSource:0}: Error finding container 1d383dc0b0d21b7ed190a52d6d55a9c12d04997a6febac24c97d76b6f496012d: Status 404 returned error can't find the container with id 1d383dc0b0d21b7ed190a52d6d55a9c12d04997a6febac24c97d76b6f496012d Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.347153 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" event={"ID":"eb82741c-91a2-49b9-b8f4-dd91167ec107","Type":"ContainerStarted","Data":"1d383dc0b0d21b7ed190a52d6d55a9c12d04997a6febac24c97d76b6f496012d"} Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.349097 4750 generic.go:334] "Generic (PLEG): container finished" podID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerID="97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426" exitCode=0 Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.349169 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvzmg" event={"ID":"ab00ccba-a824-42a2-a9fc-43139bcf9a0e","Type":"ContainerDied","Data":"97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426"} Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.352043 4750 generic.go:334] "Generic (PLEG): container finished" podID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerID="94d99a3ce55e2066f4fc011ca1dfeb2e5956008906bea120f4cd412615f97e87" exitCode=0 Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.352128 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v879l" event={"ID":"b880ae0c-0f6f-487b-8927-77437bb8a638","Type":"ContainerDied","Data":"94d99a3ce55e2066f4fc011ca1dfeb2e5956008906bea120f4cd412615f97e87"} Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.356966 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" event={"ID":"5872c77f-c7ee-4c4c-87c1-98fccae853c8","Type":"ContainerStarted","Data":"7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452"} Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.357011 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" event={"ID":"5872c77f-c7ee-4c4c-87c1-98fccae853c8","Type":"ContainerStarted","Data":"494e02fa1d1f4d50951e78fda6b755b898863b96c9249a53bf7936ee587653d3"} Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.360249 4750 generic.go:334] "Generic (PLEG): container finished" podID="41608683-d508-4374-a570-0e93ca339ffb" containerID="8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585" exitCode=0 Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.360301 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rj4zw" event={"ID":"41608683-d508-4374-a570-0e93ca339ffb","Type":"ContainerDied","Data":"8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585"} Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.381671 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb06a6d-cec1-41f8-9c38-bffd1eadd6b7" path="/var/lib/kubelet/pods/abb06a6d-cec1-41f8-9c38-bffd1eadd6b7/volumes" Mar 09 18:30:27 crc kubenswrapper[4750]: I0309 18:30:27.382653 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb6df7f3-70c7-4913-b305-be30b8f2a48b" path="/var/lib/kubelet/pods/eb6df7f3-70c7-4913-b305-be30b8f2a48b/volumes" Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.368949 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" event={"ID":"eb82741c-91a2-49b9-b8f4-dd91167ec107","Type":"ContainerStarted","Data":"8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685"} Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.369420 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.373253 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxwd2" event={"ID":"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082","Type":"ContainerStarted","Data":"580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e"} Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.373401 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.375519 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.377114 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.393410 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" podStartSLOduration=4.393390854 podStartE2EDuration="4.393390854s" podCreationTimestamp="2026-03-09 18:30:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:28.39176186 +0000 UTC m=+309.734234258" watchObservedRunningTime="2026-03-09 18:30:28.393390854 +0000 UTC m=+309.735863252" Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.419133 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lxwd2" podStartSLOduration=4.280545789 podStartE2EDuration="55.419113307s" podCreationTimestamp="2026-03-09 18:29:33 +0000 UTC" firstStartedPulling="2026-03-09 18:29:36.601979102 +0000 UTC m=+257.944451490" lastFinishedPulling="2026-03-09 18:30:27.74054661 +0000 UTC m=+309.083019008" observedRunningTime="2026-03-09 18:30:28.413137804 +0000 UTC m=+309.755610202" watchObservedRunningTime="2026-03-09 18:30:28.419113307 +0000 UTC m=+309.761585705" Mar 09 18:30:28 crc kubenswrapper[4750]: I0309 18:30:28.483835 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" podStartSLOduration=4.4838075459999995 podStartE2EDuration="4.483807546s" podCreationTimestamp="2026-03-09 18:30:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:28.478184933 +0000 UTC m=+309.820657331" watchObservedRunningTime="2026-03-09 18:30:28.483807546 +0000 UTC m=+309.826279944" Mar 09 18:30:29 crc kubenswrapper[4750]: I0309 18:30:29.387969 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvzmg" event={"ID":"ab00ccba-a824-42a2-a9fc-43139bcf9a0e","Type":"ContainerStarted","Data":"f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67"} Mar 09 18:30:29 crc kubenswrapper[4750]: I0309 18:30:29.393277 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v879l" event={"ID":"b880ae0c-0f6f-487b-8927-77437bb8a638","Type":"ContainerStarted","Data":"2871a74960ff306d958a4ad4d43b9c6bcc421163ffeea571672559f6268e474f"} Mar 09 18:30:29 crc kubenswrapper[4750]: I0309 18:30:29.395696 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rj4zw" event={"ID":"41608683-d508-4374-a570-0e93ca339ffb","Type":"ContainerStarted","Data":"927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7"} Mar 09 18:30:29 crc kubenswrapper[4750]: I0309 18:30:29.415537 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zvzmg" podStartSLOduration=3.251230177 podStartE2EDuration="58.415516413s" podCreationTimestamp="2026-03-09 18:29:31 +0000 UTC" firstStartedPulling="2026-03-09 18:29:33.31841194 +0000 UTC m=+254.660884338" lastFinishedPulling="2026-03-09 18:30:28.482698176 +0000 UTC m=+309.825170574" observedRunningTime="2026-03-09 18:30:29.412942073 +0000 UTC m=+310.755414481" watchObservedRunningTime="2026-03-09 18:30:29.415516413 +0000 UTC m=+310.757988811" Mar 09 18:30:29 crc kubenswrapper[4750]: I0309 18:30:29.433205 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v879l" podStartSLOduration=3.524780991 podStartE2EDuration="56.433183586s" podCreationTimestamp="2026-03-09 18:29:33 +0000 UTC" firstStartedPulling="2026-03-09 18:29:35.508695878 +0000 UTC m=+256.851168276" lastFinishedPulling="2026-03-09 18:30:28.417098473 +0000 UTC m=+309.759570871" observedRunningTime="2026-03-09 18:30:29.431311265 +0000 UTC m=+310.773783673" watchObservedRunningTime="2026-03-09 18:30:29.433183586 +0000 UTC m=+310.775655984" Mar 09 18:30:29 crc kubenswrapper[4750]: I0309 18:30:29.459529 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rj4zw" podStartSLOduration=4.214191265 podStartE2EDuration="59.459504415s" podCreationTimestamp="2026-03-09 18:29:30 +0000 UTC" firstStartedPulling="2026-03-09 18:29:33.2383082 +0000 UTC m=+254.580780598" lastFinishedPulling="2026-03-09 18:30:28.48362136 +0000 UTC m=+309.826093748" observedRunningTime="2026-03-09 18:30:29.459173656 +0000 UTC m=+310.801646064" watchObservedRunningTime="2026-03-09 18:30:29.459504415 +0000 UTC m=+310.801976813" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.242260 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.242731 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.291442 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.468587 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.468681 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.511741 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.595681 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.595796 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.648178 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.900320 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.900408 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:30:31 crc kubenswrapper[4750]: I0309 18:30:31.947622 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:30:32 crc kubenswrapper[4750]: I0309 18:30:32.458303 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:30:32 crc kubenswrapper[4750]: I0309 18:30:32.468851 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:30:33 crc kubenswrapper[4750]: I0309 18:30:33.621764 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:30:33 crc kubenswrapper[4750]: I0309 18:30:33.621865 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:30:33 crc kubenswrapper[4750]: I0309 18:30:33.665307 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:30:34 crc kubenswrapper[4750]: I0309 18:30:34.395790 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:30:34 crc kubenswrapper[4750]: I0309 18:30:34.397378 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:30:34 crc kubenswrapper[4750]: I0309 18:30:34.490223 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:30:34 crc kubenswrapper[4750]: I0309 18:30:34.833701 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:30:34 crc kubenswrapper[4750]: I0309 18:30:34.833779 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:30:34 crc kubenswrapper[4750]: I0309 18:30:34.878185 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:30:35 crc kubenswrapper[4750]: I0309 18:30:35.457093 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lxwd2" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="registry-server" probeResult="failure" output=< Mar 09 18:30:35 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 18:30:35 crc kubenswrapper[4750]: > Mar 09 18:30:35 crc kubenswrapper[4750]: I0309 18:30:35.505545 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:30:35 crc kubenswrapper[4750]: I0309 18:30:35.608624 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86fld"] Mar 09 18:30:35 crc kubenswrapper[4750]: I0309 18:30:35.608954 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-86fld" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="registry-server" containerID="cri-o://2923e088f5d0b051431ec290da04e6ee201bdfd710e3b95ebc8b3a151d27bd97" gracePeriod=2 Mar 09 18:30:35 crc kubenswrapper[4750]: I0309 18:30:35.808866 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dk5bm"] Mar 09 18:30:35 crc kubenswrapper[4750]: I0309 18:30:35.809182 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dk5bm" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="registry-server" containerID="cri-o://2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7" gracePeriod=2 Mar 09 18:30:36 crc kubenswrapper[4750]: I0309 18:30:36.466557 4750 generic.go:334] "Generic (PLEG): container finished" podID="3418a0da-42b0-4ac4-8957-87ed04209742" containerID="2923e088f5d0b051431ec290da04e6ee201bdfd710e3b95ebc8b3a151d27bd97" exitCode=0 Mar 09 18:30:36 crc kubenswrapper[4750]: I0309 18:30:36.466649 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86fld" event={"ID":"3418a0da-42b0-4ac4-8957-87ed04209742","Type":"ContainerDied","Data":"2923e088f5d0b051431ec290da04e6ee201bdfd710e3b95ebc8b3a151d27bd97"} Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.229947 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.350164 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.365321 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-utilities\") pod \"3418a0da-42b0-4ac4-8957-87ed04209742\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.365453 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-catalog-content\") pod \"3418a0da-42b0-4ac4-8957-87ed04209742\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.365521 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ns2k\" (UniqueName: \"kubernetes.io/projected/3418a0da-42b0-4ac4-8957-87ed04209742-kube-api-access-2ns2k\") pod \"3418a0da-42b0-4ac4-8957-87ed04209742\" (UID: \"3418a0da-42b0-4ac4-8957-87ed04209742\") " Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.367270 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-utilities" (OuterVolumeSpecName: "utilities") pod "3418a0da-42b0-4ac4-8957-87ed04209742" (UID: "3418a0da-42b0-4ac4-8957-87ed04209742"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.374322 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3418a0da-42b0-4ac4-8957-87ed04209742-kube-api-access-2ns2k" (OuterVolumeSpecName: "kube-api-access-2ns2k") pod "3418a0da-42b0-4ac4-8957-87ed04209742" (UID: "3418a0da-42b0-4ac4-8957-87ed04209742"). InnerVolumeSpecName "kube-api-access-2ns2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.425228 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3418a0da-42b0-4ac4-8957-87ed04209742" (UID: "3418a0da-42b0-4ac4-8957-87ed04209742"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.466764 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-utilities\") pod \"7c78295a-888f-4690-b970-0e1557de4ad5\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.466891 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-catalog-content\") pod \"7c78295a-888f-4690-b970-0e1557de4ad5\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.467033 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c66jx\" (UniqueName: \"kubernetes.io/projected/7c78295a-888f-4690-b970-0e1557de4ad5-kube-api-access-c66jx\") pod \"7c78295a-888f-4690-b970-0e1557de4ad5\" (UID: \"7c78295a-888f-4690-b970-0e1557de4ad5\") " Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.467419 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ns2k\" (UniqueName: \"kubernetes.io/projected/3418a0da-42b0-4ac4-8957-87ed04209742-kube-api-access-2ns2k\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.467450 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.467470 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3418a0da-42b0-4ac4-8957-87ed04209742-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.469132 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-utilities" (OuterVolumeSpecName: "utilities") pod "7c78295a-888f-4690-b970-0e1557de4ad5" (UID: "7c78295a-888f-4690-b970-0e1557de4ad5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.472433 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c78295a-888f-4690-b970-0e1557de4ad5-kube-api-access-c66jx" (OuterVolumeSpecName: "kube-api-access-c66jx") pod "7c78295a-888f-4690-b970-0e1557de4ad5" (UID: "7c78295a-888f-4690-b970-0e1557de4ad5"). InnerVolumeSpecName "kube-api-access-c66jx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.477520 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86fld" event={"ID":"3418a0da-42b0-4ac4-8957-87ed04209742","Type":"ContainerDied","Data":"ae94e61e466a995bad1ca251a6e3728b05948efb2bac95213dcc4da95ecea0cd"} Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.477604 4750 scope.go:117] "RemoveContainer" containerID="2923e088f5d0b051431ec290da04e6ee201bdfd710e3b95ebc8b3a151d27bd97" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.477544 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86fld" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.481781 4750 generic.go:334] "Generic (PLEG): container finished" podID="7c78295a-888f-4690-b970-0e1557de4ad5" containerID="2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7" exitCode=0 Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.481908 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk5bm" event={"ID":"7c78295a-888f-4690-b970-0e1557de4ad5","Type":"ContainerDied","Data":"2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7"} Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.481955 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk5bm" event={"ID":"7c78295a-888f-4690-b970-0e1557de4ad5","Type":"ContainerDied","Data":"2ebb33a38039631a51c4c042920f0ebe9aebc375df85a06ec1041d607cb4a2d7"} Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.482464 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk5bm" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.506867 4750 scope.go:117] "RemoveContainer" containerID="2b4f956de617c1de2bb545878d101e514a90c15330b64f5836c7a68b098f0d42" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.524220 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86fld"] Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.526761 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-86fld"] Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.532708 4750 scope.go:117] "RemoveContainer" containerID="6bf37f8d34bfc259388ee223d5327a75cb31a178b9fa3eef0c76d2032b060020" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.548745 4750 scope.go:117] "RemoveContainer" containerID="2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.558055 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c78295a-888f-4690-b970-0e1557de4ad5" (UID: "7c78295a-888f-4690-b970-0e1557de4ad5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.567898 4750 scope.go:117] "RemoveContainer" containerID="80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.568806 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c66jx\" (UniqueName: \"kubernetes.io/projected/7c78295a-888f-4690-b970-0e1557de4ad5-kube-api-access-c66jx\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.568848 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.568864 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c78295a-888f-4690-b970-0e1557de4ad5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.586973 4750 scope.go:117] "RemoveContainer" containerID="a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.605237 4750 scope.go:117] "RemoveContainer" containerID="2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7" Mar 09 18:30:37 crc kubenswrapper[4750]: E0309 18:30:37.605879 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7\": container with ID starting with 2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7 not found: ID does not exist" containerID="2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.605945 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7"} err="failed to get container status \"2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7\": rpc error: code = NotFound desc = could not find container \"2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7\": container with ID starting with 2fb1bd7ebad78457c1201e3f722013e5f9b5c4508551c238018b3101219cf0c7 not found: ID does not exist" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.605979 4750 scope.go:117] "RemoveContainer" containerID="80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6" Mar 09 18:30:37 crc kubenswrapper[4750]: E0309 18:30:37.607406 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6\": container with ID starting with 80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6 not found: ID does not exist" containerID="80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.607444 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6"} err="failed to get container status \"80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6\": rpc error: code = NotFound desc = could not find container \"80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6\": container with ID starting with 80dd39a50f756c9692db3dc250fb5c383d9eb731c0773971fe802c02fc1476d6 not found: ID does not exist" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.607466 4750 scope.go:117] "RemoveContainer" containerID="a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2" Mar 09 18:30:37 crc kubenswrapper[4750]: E0309 18:30:37.607851 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2\": container with ID starting with a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2 not found: ID does not exist" containerID="a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.607873 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2"} err="failed to get container status \"a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2\": rpc error: code = NotFound desc = could not find container \"a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2\": container with ID starting with a2283730d2f6815ae63ac42b3cad54194e3a5f896f27ec08356ee419a9ffc4f2 not found: ID does not exist" Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.833606 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dk5bm"] Mar 09 18:30:37 crc kubenswrapper[4750]: I0309 18:30:37.842543 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dk5bm"] Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.008168 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v879l"] Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.008425 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v879l" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="registry-server" containerID="cri-o://2871a74960ff306d958a4ad4d43b9c6bcc421163ffeea571672559f6268e474f" gracePeriod=2 Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.207271 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-76nqh"] Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.207566 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-76nqh" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="registry-server" containerID="cri-o://2e7fdef072a9836002dbfdbce3591c960bf426ed1c793fe197e966459a1f1c6a" gracePeriod=2 Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.489818 4750 generic.go:334] "Generic (PLEG): container finished" podID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerID="2871a74960ff306d958a4ad4d43b9c6bcc421163ffeea571672559f6268e474f" exitCode=0 Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.489889 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v879l" event={"ID":"b880ae0c-0f6f-487b-8927-77437bb8a638","Type":"ContainerDied","Data":"2871a74960ff306d958a4ad4d43b9c6bcc421163ffeea571672559f6268e474f"} Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.494223 4750 generic.go:334] "Generic (PLEG): container finished" podID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerID="2e7fdef072a9836002dbfdbce3591c960bf426ed1c793fe197e966459a1f1c6a" exitCode=0 Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.494269 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76nqh" event={"ID":"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7","Type":"ContainerDied","Data":"2e7fdef072a9836002dbfdbce3591c960bf426ed1c793fe197e966459a1f1c6a"} Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.574911 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.684731 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-utilities\") pod \"b880ae0c-0f6f-487b-8927-77437bb8a638\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.684892 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-catalog-content\") pod \"b880ae0c-0f6f-487b-8927-77437bb8a638\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.684986 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k5lr\" (UniqueName: \"kubernetes.io/projected/b880ae0c-0f6f-487b-8927-77437bb8a638-kube-api-access-9k5lr\") pod \"b880ae0c-0f6f-487b-8927-77437bb8a638\" (UID: \"b880ae0c-0f6f-487b-8927-77437bb8a638\") " Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.685899 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-utilities" (OuterVolumeSpecName: "utilities") pod "b880ae0c-0f6f-487b-8927-77437bb8a638" (UID: "b880ae0c-0f6f-487b-8927-77437bb8a638"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.690085 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b880ae0c-0f6f-487b-8927-77437bb8a638-kube-api-access-9k5lr" (OuterVolumeSpecName: "kube-api-access-9k5lr") pod "b880ae0c-0f6f-487b-8927-77437bb8a638" (UID: "b880ae0c-0f6f-487b-8927-77437bb8a638"). InnerVolumeSpecName "kube-api-access-9k5lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.709447 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b880ae0c-0f6f-487b-8927-77437bb8a638" (UID: "b880ae0c-0f6f-487b-8927-77437bb8a638"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.740856 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.786851 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-catalog-content\") pod \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.786933 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-utilities\") pod \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.786964 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn6pd\" (UniqueName: \"kubernetes.io/projected/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-kube-api-access-dn6pd\") pod \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\" (UID: \"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7\") " Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.787252 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.787281 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b880ae0c-0f6f-487b-8927-77437bb8a638-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.787298 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k5lr\" (UniqueName: \"kubernetes.io/projected/b880ae0c-0f6f-487b-8927-77437bb8a638-kube-api-access-9k5lr\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.788100 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-utilities" (OuterVolumeSpecName: "utilities") pod "3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" (UID: "3c35f63c-cbc8-4d2b-b971-3e18a8222bc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.792109 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-kube-api-access-dn6pd" (OuterVolumeSpecName: "kube-api-access-dn6pd") pod "3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" (UID: "3c35f63c-cbc8-4d2b-b971-3e18a8222bc7"). InnerVolumeSpecName "kube-api-access-dn6pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.888323 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.888375 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn6pd\" (UniqueName: \"kubernetes.io/projected/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-kube-api-access-dn6pd\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.919845 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" (UID: "3c35f63c-cbc8-4d2b-b971-3e18a8222bc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:30:38 crc kubenswrapper[4750]: I0309 18:30:38.989739 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.384190 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" path="/var/lib/kubelet/pods/3418a0da-42b0-4ac4-8957-87ed04209742/volumes" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.385308 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" path="/var/lib/kubelet/pods/7c78295a-888f-4690-b970-0e1557de4ad5/volumes" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.506124 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76nqh" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.506112 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76nqh" event={"ID":"3c35f63c-cbc8-4d2b-b971-3e18a8222bc7","Type":"ContainerDied","Data":"f6413534428f3d4968e7e1b5177af59980b14d8573513ff1322749b2c048fcb6"} Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.506586 4750 scope.go:117] "RemoveContainer" containerID="2e7fdef072a9836002dbfdbce3591c960bf426ed1c793fe197e966459a1f1c6a" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.521233 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v879l" event={"ID":"b880ae0c-0f6f-487b-8927-77437bb8a638","Type":"ContainerDied","Data":"07ce79bd8479ba3c148f0eba9a93d1882bcc6e5310ce64049d3f04dc917f27ef"} Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.521376 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v879l" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.535517 4750 scope.go:117] "RemoveContainer" containerID="0afceb5bb80b11f325cf4af12963ca64b215c4d515f87e58ff50400fbb789897" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.536765 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-76nqh"] Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.542654 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-76nqh"] Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.557051 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v879l"] Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.559076 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v879l"] Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.560736 4750 scope.go:117] "RemoveContainer" containerID="e423261a174b2b891ba24702f2e32a44290a58ae99ca00baca80127fcc89780f" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.601120 4750 scope.go:117] "RemoveContainer" containerID="2871a74960ff306d958a4ad4d43b9c6bcc421163ffeea571672559f6268e474f" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.621716 4750 scope.go:117] "RemoveContainer" containerID="94d99a3ce55e2066f4fc011ca1dfeb2e5956008906bea120f4cd412615f97e87" Mar 09 18:30:39 crc kubenswrapper[4750]: I0309 18:30:39.640089 4750 scope.go:117] "RemoveContainer" containerID="21d7ac9d0e4961907d815247e020c5032d2b2aa88b34353240f94a9753b074bc" Mar 09 18:30:41 crc kubenswrapper[4750]: I0309 18:30:41.318040 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:30:41 crc kubenswrapper[4750]: I0309 18:30:41.387686 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" path="/var/lib/kubelet/pods/3c35f63c-cbc8-4d2b-b971-3e18a8222bc7/volumes" Mar 09 18:30:41 crc kubenswrapper[4750]: I0309 18:30:41.388586 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" path="/var/lib/kubelet/pods/b880ae0c-0f6f-487b-8927-77437bb8a638/volumes" Mar 09 18:30:41 crc kubenswrapper[4750]: I0309 18:30:41.517612 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:30:44 crc kubenswrapper[4750]: I0309 18:30:44.454107 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:30:44 crc kubenswrapper[4750]: I0309 18:30:44.501409 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-859c744c7b-skzpn"] Mar 09 18:30:44 crc kubenswrapper[4750]: I0309 18:30:44.501779 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" podUID="5872c77f-c7ee-4c4c-87c1-98fccae853c8" containerName="controller-manager" containerID="cri-o://7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452" gracePeriod=30 Mar 09 18:30:44 crc kubenswrapper[4750]: I0309 18:30:44.530063 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:30:44 crc kubenswrapper[4750]: I0309 18:30:44.607145 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k"] Mar 09 18:30:44 crc kubenswrapper[4750]: I0309 18:30:44.607377 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" podUID="eb82741c-91a2-49b9-b8f4-dd91167ec107" containerName="route-controller-manager" containerID="cri-o://8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685" gracePeriod=30 Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.165985 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.170533 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.292167 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-client-ca\") pod \"eb82741c-91a2-49b9-b8f4-dd91167ec107\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.292256 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-config\") pod \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.292308 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-proxy-ca-bundles\") pod \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.292373 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqj8z\" (UniqueName: \"kubernetes.io/projected/eb82741c-91a2-49b9-b8f4-dd91167ec107-kube-api-access-dqj8z\") pod \"eb82741c-91a2-49b9-b8f4-dd91167ec107\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.292428 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5872c77f-c7ee-4c4c-87c1-98fccae853c8-serving-cert\") pod \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.292484 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb82741c-91a2-49b9-b8f4-dd91167ec107-serving-cert\") pod \"eb82741c-91a2-49b9-b8f4-dd91167ec107\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.293646 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-config" (OuterVolumeSpecName: "config") pod "5872c77f-c7ee-4c4c-87c1-98fccae853c8" (UID: "5872c77f-c7ee-4c4c-87c1-98fccae853c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.293655 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5872c77f-c7ee-4c4c-87c1-98fccae853c8" (UID: "5872c77f-c7ee-4c4c-87c1-98fccae853c8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.293871 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-client-ca" (OuterVolumeSpecName: "client-ca") pod "eb82741c-91a2-49b9-b8f4-dd91167ec107" (UID: "eb82741c-91a2-49b9-b8f4-dd91167ec107"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.294045 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-config\") pod \"eb82741c-91a2-49b9-b8f4-dd91167ec107\" (UID: \"eb82741c-91a2-49b9-b8f4-dd91167ec107\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.294194 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-client-ca\") pod \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.294240 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fjq5\" (UniqueName: \"kubernetes.io/projected/5872c77f-c7ee-4c4c-87c1-98fccae853c8-kube-api-access-6fjq5\") pod \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\" (UID: \"5872c77f-c7ee-4c4c-87c1-98fccae853c8\") " Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.294749 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.294769 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.294780 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.295276 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-config" (OuterVolumeSpecName: "config") pod "eb82741c-91a2-49b9-b8f4-dd91167ec107" (UID: "eb82741c-91a2-49b9-b8f4-dd91167ec107"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.295720 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-client-ca" (OuterVolumeSpecName: "client-ca") pod "5872c77f-c7ee-4c4c-87c1-98fccae853c8" (UID: "5872c77f-c7ee-4c4c-87c1-98fccae853c8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.300256 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb82741c-91a2-49b9-b8f4-dd91167ec107-kube-api-access-dqj8z" (OuterVolumeSpecName: "kube-api-access-dqj8z") pod "eb82741c-91a2-49b9-b8f4-dd91167ec107" (UID: "eb82741c-91a2-49b9-b8f4-dd91167ec107"). InnerVolumeSpecName "kube-api-access-dqj8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.300273 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb82741c-91a2-49b9-b8f4-dd91167ec107-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "eb82741c-91a2-49b9-b8f4-dd91167ec107" (UID: "eb82741c-91a2-49b9-b8f4-dd91167ec107"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.300371 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5872c77f-c7ee-4c4c-87c1-98fccae853c8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5872c77f-c7ee-4c4c-87c1-98fccae853c8" (UID: "5872c77f-c7ee-4c4c-87c1-98fccae853c8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.301135 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5872c77f-c7ee-4c4c-87c1-98fccae853c8-kube-api-access-6fjq5" (OuterVolumeSpecName: "kube-api-access-6fjq5") pod "5872c77f-c7ee-4c4c-87c1-98fccae853c8" (UID: "5872c77f-c7ee-4c4c-87c1-98fccae853c8"). InnerVolumeSpecName "kube-api-access-6fjq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.397275 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqj8z\" (UniqueName: \"kubernetes.io/projected/eb82741c-91a2-49b9-b8f4-dd91167ec107-kube-api-access-dqj8z\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.397808 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5872c77f-c7ee-4c4c-87c1-98fccae853c8-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.397827 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb82741c-91a2-49b9-b8f4-dd91167ec107-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.397843 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb82741c-91a2-49b9-b8f4-dd91167ec107-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.397857 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5872c77f-c7ee-4c4c-87c1-98fccae853c8-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.397905 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fjq5\" (UniqueName: \"kubernetes.io/projected/5872c77f-c7ee-4c4c-87c1-98fccae853c8-kube-api-access-6fjq5\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.579796 4750 generic.go:334] "Generic (PLEG): container finished" podID="eb82741c-91a2-49b9-b8f4-dd91167ec107" containerID="8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685" exitCode=0 Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.579916 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" event={"ID":"eb82741c-91a2-49b9-b8f4-dd91167ec107","Type":"ContainerDied","Data":"8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685"} Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.579978 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.580866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k" event={"ID":"eb82741c-91a2-49b9-b8f4-dd91167ec107","Type":"ContainerDied","Data":"1d383dc0b0d21b7ed190a52d6d55a9c12d04997a6febac24c97d76b6f496012d"} Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.580890 4750 scope.go:117] "RemoveContainer" containerID="8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.584825 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.584779 4750 generic.go:334] "Generic (PLEG): container finished" podID="5872c77f-c7ee-4c4c-87c1-98fccae853c8" containerID="7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452" exitCode=0 Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.584983 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" event={"ID":"5872c77f-c7ee-4c4c-87c1-98fccae853c8","Type":"ContainerDied","Data":"7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452"} Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.586820 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-859c744c7b-skzpn" event={"ID":"5872c77f-c7ee-4c4c-87c1-98fccae853c8","Type":"ContainerDied","Data":"494e02fa1d1f4d50951e78fda6b755b898863b96c9249a53bf7936ee587653d3"} Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.607777 4750 scope.go:117] "RemoveContainer" containerID="8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685" Mar 09 18:30:45 crc kubenswrapper[4750]: E0309 18:30:45.608354 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685\": container with ID starting with 8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685 not found: ID does not exist" containerID="8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.608399 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685"} err="failed to get container status \"8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685\": rpc error: code = NotFound desc = could not find container \"8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685\": container with ID starting with 8b7578cf9aa154900c8637eb4e80f5d2365f6c31373c943f3960acb5f0bef685 not found: ID does not exist" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.608426 4750 scope.go:117] "RemoveContainer" containerID="7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.614857 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k"] Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.621848 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7649f4cc88-6dg2k"] Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.623517 4750 scope.go:117] "RemoveContainer" containerID="7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452" Mar 09 18:30:45 crc kubenswrapper[4750]: E0309 18:30:45.623867 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452\": container with ID starting with 7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452 not found: ID does not exist" containerID="7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.623901 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452"} err="failed to get container status \"7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452\": rpc error: code = NotFound desc = could not find container \"7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452\": container with ID starting with 7f07db435ec476c2ad7dc15e53b5f8b35614d4da6a4ece805a2f8b2e2effa452 not found: ID does not exist" Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.644835 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-859c744c7b-skzpn"] Mar 09 18:30:45 crc kubenswrapper[4750]: I0309 18:30:45.650929 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-859c744c7b-skzpn"] Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399004 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c66bc864c-rcw4h"] Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399467 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5872c77f-c7ee-4c4c-87c1-98fccae853c8" containerName="controller-manager" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399490 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5872c77f-c7ee-4c4c-87c1-98fccae853c8" containerName="controller-manager" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399521 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399535 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399557 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399569 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399587 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399599 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399619 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399659 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399676 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399688 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399703 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399714 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399738 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399750 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399765 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399777 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399792 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399807 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="extract-utilities" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399822 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399833 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399852 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399866 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399879 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399891 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="extract-content" Mar 09 18:30:46 crc kubenswrapper[4750]: E0309 18:30:46.399906 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb82741c-91a2-49b9-b8f4-dd91167ec107" containerName="route-controller-manager" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.399917 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb82741c-91a2-49b9-b8f4-dd91167ec107" containerName="route-controller-manager" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.400098 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c78295a-888f-4690-b970-0e1557de4ad5" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.400122 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3418a0da-42b0-4ac4-8957-87ed04209742" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.400140 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b880ae0c-0f6f-487b-8927-77437bb8a638" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.400162 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5872c77f-c7ee-4c4c-87c1-98fccae853c8" containerName="controller-manager" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.400184 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb82741c-91a2-49b9-b8f4-dd91167ec107" containerName="route-controller-manager" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.400199 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c35f63c-cbc8-4d2b-b971-3e18a8222bc7" containerName="registry-server" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.401003 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.402299 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t"] Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.403147 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.405278 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.405800 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.405959 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.406103 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.406249 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.408021 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.408188 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.408385 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.408672 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.408818 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.409055 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.409342 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.417061 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.423983 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t"] Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.428669 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c66bc864c-rcw4h"] Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523035 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-serving-cert\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523129 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-config\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523155 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-client-ca\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523172 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-serving-cert\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523199 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzm9n\" (UniqueName: \"kubernetes.io/projected/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-kube-api-access-lzm9n\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523228 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-proxy-ca-bundles\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523249 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-client-ca\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523292 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st25d\" (UniqueName: \"kubernetes.io/projected/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-kube-api-access-st25d\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.523432 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-config\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.624432 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-proxy-ca-bundles\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.624516 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-client-ca\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.624581 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st25d\" (UniqueName: \"kubernetes.io/projected/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-kube-api-access-st25d\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.624727 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-config\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.626204 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-client-ca\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.626412 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-config\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.626484 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-serving-cert\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.626616 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-proxy-ca-bundles\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.627024 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-config\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.627051 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-client-ca\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.627072 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-serving-cert\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.627091 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzm9n\" (UniqueName: \"kubernetes.io/projected/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-kube-api-access-lzm9n\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.628019 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-client-ca\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.628528 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-config\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.638950 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-serving-cert\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.640332 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-serving-cert\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.644816 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st25d\" (UniqueName: \"kubernetes.io/projected/9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7-kube-api-access-st25d\") pod \"controller-manager-c66bc864c-rcw4h\" (UID: \"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7\") " pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.657062 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzm9n\" (UniqueName: \"kubernetes.io/projected/f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1-kube-api-access-lzm9n\") pod \"route-controller-manager-677547cf69-rqx5t\" (UID: \"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1\") " pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.727967 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.741146 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.953573 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t"] Mar 09 18:30:46 crc kubenswrapper[4750]: W0309 18:30:46.959228 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1e3f9a5_464c_4fbd_baa8_c1b148f21cd1.slice/crio-1eb5606a722521836781fba1bed94d4aa1595efc9e18856a25141012ed3635bb WatchSource:0}: Error finding container 1eb5606a722521836781fba1bed94d4aa1595efc9e18856a25141012ed3635bb: Status 404 returned error can't find the container with id 1eb5606a722521836781fba1bed94d4aa1595efc9e18856a25141012ed3635bb Mar 09 18:30:46 crc kubenswrapper[4750]: I0309 18:30:46.996964 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c66bc864c-rcw4h"] Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.384289 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5872c77f-c7ee-4c4c-87c1-98fccae853c8" path="/var/lib/kubelet/pods/5872c77f-c7ee-4c4c-87c1-98fccae853c8/volumes" Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.385657 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb82741c-91a2-49b9-b8f4-dd91167ec107" path="/var/lib/kubelet/pods/eb82741c-91a2-49b9-b8f4-dd91167ec107/volumes" Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.606398 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" event={"ID":"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7","Type":"ContainerStarted","Data":"acdeaeccb9b775c187d2e53ed0a8433016bcd43bc6ba6bff51118046a96c2331"} Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.606980 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.607143 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" event={"ID":"9e0b5e4d-9f4b-45c2-a0de-0a0daa8f85a7","Type":"ContainerStarted","Data":"8ec7a2c5ceec9321eb880d65bc03b4270ab2d8af2c51fea17881e0ac1865bd40"} Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.610219 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" event={"ID":"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1","Type":"ContainerStarted","Data":"610b7e92499db85d39906c9672514a1560ab519abef14024a0dbf2939d86d33b"} Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.610279 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" event={"ID":"f1e3f9a5-464c-4fbd-baa8-c1b148f21cd1","Type":"ContainerStarted","Data":"1eb5606a722521836781fba1bed94d4aa1595efc9e18856a25141012ed3635bb"} Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.610545 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.618343 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.629323 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c66bc864c-rcw4h" podStartSLOduration=3.62929057 podStartE2EDuration="3.62929057s" podCreationTimestamp="2026-03-09 18:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:47.624726072 +0000 UTC m=+328.967198500" watchObservedRunningTime="2026-03-09 18:30:47.62929057 +0000 UTC m=+328.971763008" Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.678110 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" podStartSLOduration=3.678040892 podStartE2EDuration="3.678040892s" podCreationTimestamp="2026-03-09 18:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:30:47.676061787 +0000 UTC m=+329.018534225" watchObservedRunningTime="2026-03-09 18:30:47.678040892 +0000 UTC m=+329.020513330" Mar 09 18:30:47 crc kubenswrapper[4750]: I0309 18:30:47.742665 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-677547cf69-rqx5t" Mar 09 18:30:51 crc kubenswrapper[4750]: I0309 18:30:51.093700 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jcrcb"] Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.365792 4750 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.367092 4750 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.367262 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.367561 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad" gracePeriod=15 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.367645 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac" gracePeriod=15 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.367676 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334" gracePeriod=15 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.367727 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c" gracePeriod=15 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.367569 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14" gracePeriod=15 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.369984 4750 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370196 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370211 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370220 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370227 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370236 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370243 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370255 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370262 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370269 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370274 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370284 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370290 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370299 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370307 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370326 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370332 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370341 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370349 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370465 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370478 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370489 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370499 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370508 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370522 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370532 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370643 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370652 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.370667 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370673 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370783 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370794 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.370802 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.433374 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457010 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457091 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457191 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457216 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457440 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457501 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457527 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.457573 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.559607 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.559680 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.559799 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.559857 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.559903 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.559934 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.559978 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560036 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560106 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560106 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560157 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560183 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560249 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560279 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560280 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.560358 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.653717 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.655369 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.656614 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14" exitCode=0 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.656681 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334" exitCode=0 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.656703 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac" exitCode=0 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.656730 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c" exitCode=2 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.656724 4750 scope.go:117] "RemoveContainer" containerID="c2323967706e64ad37d7dae311f9be9f23abd591eb73cba11c976e35d673d3ba" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.664594 4750 generic.go:334] "Generic (PLEG): container finished" podID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" containerID="48f6a8232e40d5d679843d5bc884eae21d8c36abeb7755e85aa03dfb1ede06cf" exitCode=0 Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.664673 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243","Type":"ContainerDied","Data":"48f6a8232e40d5d679843d5bc884eae21d8c36abeb7755e85aa03dfb1ede06cf"} Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.666084 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.666593 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:53 crc kubenswrapper[4750]: I0309 18:30:53.734551 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:30:53 crc kubenswrapper[4750]: E0309 18:30:53.769186 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.93:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189b3fd2ffd63d7f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:30:53.768334719 +0000 UTC m=+335.110807117,LastTimestamp:2026-03-09 18:30:53.768334719 +0000 UTC m=+335.110807117,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:30:54 crc kubenswrapper[4750]: I0309 18:30:54.672170 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb"} Mar 09 18:30:54 crc kubenswrapper[4750]: I0309 18:30:54.672714 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"43b84eb732eca49731c441e73329bbe7d7d83dc757390b5ce930a0db0fe84e32"} Mar 09 18:30:54 crc kubenswrapper[4750]: I0309 18:30:54.673899 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:54 crc kubenswrapper[4750]: I0309 18:30:54.674133 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:54 crc kubenswrapper[4750]: I0309 18:30:54.676393 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.130107 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.131784 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.132761 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.286298 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-var-lock\") pod \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.286387 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kubelet-dir\") pod \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.286441 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-var-lock" (OuterVolumeSpecName: "var-lock") pod "8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" (UID: "8e321a00-cd51-4fc5-b20f-8a6c5d7f3243"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.286516 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kube-api-access\") pod \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\" (UID: \"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243\") " Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.286539 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" (UID: "8e321a00-cd51-4fc5-b20f-8a6c5d7f3243"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.286853 4750 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-var-lock\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.286874 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.292309 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" (UID: "8e321a00-cd51-4fc5-b20f-8a6c5d7f3243"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.388610 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e321a00-cd51-4fc5-b20f-8a6c5d7f3243-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.686152 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.686136 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8e321a00-cd51-4fc5-b20f-8a6c5d7f3243","Type":"ContainerDied","Data":"e0edc50f0132806189b063abe684e4b58ff151e63c4ec71a441b986483858788"} Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.687797 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0edc50f0132806189b063abe684e4b58ff151e63c4ec71a441b986483858788" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.690880 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.691408 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.842319 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.843582 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.844401 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.844870 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.845329 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999119 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999201 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999253 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999338 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999408 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999498 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999730 4750 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999755 4750 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:55 crc kubenswrapper[4750]: I0309 18:30:55.999774 4750 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.698966 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.700407 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad" exitCode=0 Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.700508 4750 scope.go:117] "RemoveContainer" containerID="afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.700482 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.724797 4750 scope.go:117] "RemoveContainer" containerID="66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.725874 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.726557 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.727160 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.750509 4750 scope.go:117] "RemoveContainer" containerID="249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.768515 4750 scope.go:117] "RemoveContainer" containerID="832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.787706 4750 scope.go:117] "RemoveContainer" containerID="b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.806597 4750 scope.go:117] "RemoveContainer" containerID="5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.830836 4750 scope.go:117] "RemoveContainer" containerID="afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14" Mar 09 18:30:56 crc kubenswrapper[4750]: E0309 18:30:56.831504 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14\": container with ID starting with afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14 not found: ID does not exist" containerID="afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.831554 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14"} err="failed to get container status \"afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14\": rpc error: code = NotFound desc = could not find container \"afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14\": container with ID starting with afd132f64b4472253b14497fe46ab768b6fd26edaa008fe85c2b08b2c2724d14 not found: ID does not exist" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.831578 4750 scope.go:117] "RemoveContainer" containerID="66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334" Mar 09 18:30:56 crc kubenswrapper[4750]: E0309 18:30:56.832767 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\": container with ID starting with 66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334 not found: ID does not exist" containerID="66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.832830 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334"} err="failed to get container status \"66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\": rpc error: code = NotFound desc = could not find container \"66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334\": container with ID starting with 66c995bb51f310bd1de49d93c7db314e00ef3cb6e9b47c38a1960009a6366334 not found: ID does not exist" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.832868 4750 scope.go:117] "RemoveContainer" containerID="249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac" Mar 09 18:30:56 crc kubenswrapper[4750]: E0309 18:30:56.833211 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\": container with ID starting with 249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac not found: ID does not exist" containerID="249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.833240 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac"} err="failed to get container status \"249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\": rpc error: code = NotFound desc = could not find container \"249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac\": container with ID starting with 249fdb399d09221bb71fbda7506921cace99be8b74bb7c38ea0e833b1c5598ac not found: ID does not exist" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.833259 4750 scope.go:117] "RemoveContainer" containerID="832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c" Mar 09 18:30:56 crc kubenswrapper[4750]: E0309 18:30:56.833556 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\": container with ID starting with 832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c not found: ID does not exist" containerID="832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.833589 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c"} err="failed to get container status \"832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\": rpc error: code = NotFound desc = could not find container \"832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c\": container with ID starting with 832d15f131a32fcf442487bcd86a049b535869b9c2ccfa60227a883d671caa8c not found: ID does not exist" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.833609 4750 scope.go:117] "RemoveContainer" containerID="b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad" Mar 09 18:30:56 crc kubenswrapper[4750]: E0309 18:30:56.833856 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\": container with ID starting with b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad not found: ID does not exist" containerID="b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.833938 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad"} err="failed to get container status \"b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\": rpc error: code = NotFound desc = could not find container \"b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad\": container with ID starting with b70664670e7ae6bb4e977ab7634b13fb356bdbae0da07bae2561f0fa7ae2abad not found: ID does not exist" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.833955 4750 scope.go:117] "RemoveContainer" containerID="5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3" Mar 09 18:30:56 crc kubenswrapper[4750]: E0309 18:30:56.834273 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\": container with ID starting with 5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3 not found: ID does not exist" containerID="5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3" Mar 09 18:30:56 crc kubenswrapper[4750]: I0309 18:30:56.834300 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3"} err="failed to get container status \"5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\": rpc error: code = NotFound desc = could not find container \"5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3\": container with ID starting with 5f67cc75acb2867b40bb50d13bdeb64643147044f729b3d76c79f1c5982f01f3 not found: ID does not exist" Mar 09 18:30:56 crc kubenswrapper[4750]: E0309 18:30:56.982168 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.93:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189b3fd2ffd63d7f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:30:53.768334719 +0000 UTC m=+335.110807117,LastTimestamp:2026-03-09 18:30:53.768334719 +0000 UTC m=+335.110807117,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:30:57 crc kubenswrapper[4750]: I0309 18:30:57.384568 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 09 18:30:57 crc kubenswrapper[4750]: E0309 18:30:57.452999 4750 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.93:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" volumeName="registry-storage" Mar 09 18:30:59 crc kubenswrapper[4750]: I0309 18:30:59.380751 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:30:59 crc kubenswrapper[4750]: I0309 18:30:59.384421 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:01 crc kubenswrapper[4750]: E0309 18:31:01.255278 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:31:01Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:31:01Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:31:01Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T18:31:01Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:01 crc kubenswrapper[4750]: E0309 18:31:01.255900 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:01 crc kubenswrapper[4750]: E0309 18:31:01.256968 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:01 crc kubenswrapper[4750]: E0309 18:31:01.257315 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:01 crc kubenswrapper[4750]: E0309 18:31:01.257488 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:01 crc kubenswrapper[4750]: E0309 18:31:01.257499 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 18:31:02 crc kubenswrapper[4750]: E0309 18:31:02.556985 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:02 crc kubenswrapper[4750]: E0309 18:31:02.558338 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:02 crc kubenswrapper[4750]: E0309 18:31:02.558909 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:02 crc kubenswrapper[4750]: E0309 18:31:02.559509 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:02 crc kubenswrapper[4750]: E0309 18:31:02.560050 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:02 crc kubenswrapper[4750]: I0309 18:31:02.560112 4750 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 09 18:31:02 crc kubenswrapper[4750]: E0309 18:31:02.560571 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="200ms" Mar 09 18:31:02 crc kubenswrapper[4750]: E0309 18:31:02.762277 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="400ms" Mar 09 18:31:03 crc kubenswrapper[4750]: E0309 18:31:03.172405 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="800ms" Mar 09 18:31:03 crc kubenswrapper[4750]: E0309 18:31:03.973522 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="1.6s" Mar 09 18:31:05 crc kubenswrapper[4750]: E0309 18:31:05.574349 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.93:6443: connect: connection refused" interval="3.2s" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.373056 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.375802 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.382064 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.400325 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.400397 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:06 crc kubenswrapper[4750]: E0309 18:31:06.401449 4750 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.402395 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:06 crc kubenswrapper[4750]: W0309 18:31:06.430665 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-d49161e254fd1f879d6e117ce3002e5dd61c5835991f8c20418b1c5a47abef2b WatchSource:0}: Error finding container d49161e254fd1f879d6e117ce3002e5dd61c5835991f8c20418b1c5a47abef2b: Status 404 returned error can't find the container with id d49161e254fd1f879d6e117ce3002e5dd61c5835991f8c20418b1c5a47abef2b Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.784290 4750 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="dfb3d34934220660e85eb8d1cc8f6d8bfa5fa88653b8733d6b5768a3805513a6" exitCode=0 Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.784438 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"dfb3d34934220660e85eb8d1cc8f6d8bfa5fa88653b8733d6b5768a3805513a6"} Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.784611 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d49161e254fd1f879d6e117ce3002e5dd61c5835991f8c20418b1c5a47abef2b"} Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.784878 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.784892 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.785403 4750 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:06 crc kubenswrapper[4750]: I0309 18:31:06.785670 4750 status_manager.go:851] "Failed to get status for pod" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" Mar 09 18:31:06 crc kubenswrapper[4750]: E0309 18:31:06.785684 4750 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.93:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:06 crc kubenswrapper[4750]: E0309 18:31:06.983170 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.93:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189b3fd2ffd63d7f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 18:30:53.768334719 +0000 UTC m=+335.110807117,LastTimestamp:2026-03-09 18:30:53.768334719 +0000 UTC m=+335.110807117,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 18:31:07 crc kubenswrapper[4750]: I0309 18:31:07.795615 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"80cec09b1686cf0f32066ea0724dfa866749afa984a7a6d54abeff7fa31402f1"} Mar 09 18:31:07 crc kubenswrapper[4750]: I0309 18:31:07.796045 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"74eb9fffdb67c6c321757dae7f990035ccdd45759c01c9abaf5a4c6ef810aa61"} Mar 09 18:31:07 crc kubenswrapper[4750]: I0309 18:31:07.796055 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"830fa687b672880d6dc967f5c966a30db277b0d6582f7ac710f4b857832abcbf"} Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.013226 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.013318 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.804066 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.805181 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.805253 4750 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940" exitCode=1 Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.805338 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940"} Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.806235 4750 scope.go:117] "RemoveContainer" containerID="be285f0dbe4ea25d0424e9276191663b28793d56f87d5013070bf34ab59a9940" Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.809502 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"509a1a631f66fd2646a2eb9fddd6b27232cdd63f243515b94ea0349e49b6b547"} Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.809554 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5e3f98f936bdee3c014b82bdc58148ceb578e4f8f28d69f14ff44d6f819c9aa4"} Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.809789 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.809886 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:08 crc kubenswrapper[4750]: I0309 18:31:08.809903 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:09 crc kubenswrapper[4750]: I0309 18:31:09.482489 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:31:09 crc kubenswrapper[4750]: I0309 18:31:09.817812 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 18:31:09 crc kubenswrapper[4750]: I0309 18:31:09.819353 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 09 18:31:09 crc kubenswrapper[4750]: I0309 18:31:09.819408 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"81ab25001753478b8959aca3107f19ced002ddf59bfa64758d79f37d0351f617"} Mar 09 18:31:11 crc kubenswrapper[4750]: I0309 18:31:11.403338 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:11 crc kubenswrapper[4750]: I0309 18:31:11.403427 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:11 crc kubenswrapper[4750]: I0309 18:31:11.409401 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:12 crc kubenswrapper[4750]: I0309 18:31:12.410477 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:31:13 crc kubenswrapper[4750]: I0309 18:31:13.817999 4750 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:13 crc kubenswrapper[4750]: I0309 18:31:13.859343 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:13 crc kubenswrapper[4750]: I0309 18:31:13.859381 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:13 crc kubenswrapper[4750]: I0309 18:31:13.863082 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:13 crc kubenswrapper[4750]: I0309 18:31:13.897178 4750 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4dbd0670-fb05-4ca4-be8b-33825349d55d" Mar 09 18:31:14 crc kubenswrapper[4750]: I0309 18:31:14.867097 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:14 crc kubenswrapper[4750]: I0309 18:31:14.867156 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee55d4c6-02a9-43f2-86f1-e5ef4faa01c9" Mar 09 18:31:14 crc kubenswrapper[4750]: I0309 18:31:14.871554 4750 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4dbd0670-fb05-4ca4-be8b-33825349d55d" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.136360 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" podUID="f821a876-eab1-4298-b06c-9560fb385085" containerName="oauth-openshift" containerID="cri-o://26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096" gracePeriod=15 Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.663989 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803529 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-provider-selection\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803603 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-router-certs\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803696 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-service-ca\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803742 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-trusted-ca-bundle\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803796 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-error\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803836 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-cliconfig\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803906 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-audit-policies\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803944 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f821a876-eab1-4298-b06c-9560fb385085-audit-dir\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.803996 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-ocp-branding-template\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.804042 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-idp-0-file-data\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.804080 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9d4l\" (UniqueName: \"kubernetes.io/projected/f821a876-eab1-4298-b06c-9560fb385085-kube-api-access-f9d4l\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.804118 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-session\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.804169 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-login\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.804199 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-serving-cert\") pod \"f821a876-eab1-4298-b06c-9560fb385085\" (UID: \"f821a876-eab1-4298-b06c-9560fb385085\") " Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.804866 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.804991 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f821a876-eab1-4298-b06c-9560fb385085-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.806545 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.806900 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.807604 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.813049 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.813170 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f821a876-eab1-4298-b06c-9560fb385085-kube-api-access-f9d4l" (OuterVolumeSpecName: "kube-api-access-f9d4l") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "kube-api-access-f9d4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.813456 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.813953 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.814125 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.815531 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.816021 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.816473 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.816673 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f821a876-eab1-4298-b06c-9560fb385085" (UID: "f821a876-eab1-4298-b06c-9560fb385085"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.886496 4750 generic.go:334] "Generic (PLEG): container finished" podID="f821a876-eab1-4298-b06c-9560fb385085" containerID="26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096" exitCode=0 Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.886561 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" event={"ID":"f821a876-eab1-4298-b06c-9560fb385085","Type":"ContainerDied","Data":"26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096"} Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.886610 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.886674 4750 scope.go:117] "RemoveContainer" containerID="26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.886624 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jcrcb" event={"ID":"f821a876-eab1-4298-b06c-9560fb385085","Type":"ContainerDied","Data":"12dca8646a722a54a0300fa6928c9ce622ac8d87b754af1c3c6c3a386ff967a5"} Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905650 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905693 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905707 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905719 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905734 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905746 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905758 4750 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f821a876-eab1-4298-b06c-9560fb385085-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905770 4750 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f821a876-eab1-4298-b06c-9560fb385085-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905783 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905797 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905809 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9d4l\" (UniqueName: \"kubernetes.io/projected/f821a876-eab1-4298-b06c-9560fb385085-kube-api-access-f9d4l\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905825 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905838 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.905852 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f821a876-eab1-4298-b06c-9560fb385085-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.909147 4750 scope.go:117] "RemoveContainer" containerID="26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096" Mar 09 18:31:16 crc kubenswrapper[4750]: E0309 18:31:16.910787 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096\": container with ID starting with 26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096 not found: ID does not exist" containerID="26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096" Mar 09 18:31:16 crc kubenswrapper[4750]: I0309 18:31:16.910860 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096"} err="failed to get container status \"26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096\": rpc error: code = NotFound desc = could not find container \"26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096\": container with ID starting with 26906f86037e5dc7cae87ab9b6ed8fb5d187be6469c7c3acbbf69d4aee966096 not found: ID does not exist" Mar 09 18:31:19 crc kubenswrapper[4750]: I0309 18:31:19.482349 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:31:19 crc kubenswrapper[4750]: I0309 18:31:19.487798 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:31:19 crc kubenswrapper[4750]: I0309 18:31:19.913146 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 18:31:23 crc kubenswrapper[4750]: I0309 18:31:23.271088 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 09 18:31:23 crc kubenswrapper[4750]: I0309 18:31:23.599390 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 09 18:31:23 crc kubenswrapper[4750]: I0309 18:31:23.767040 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 09 18:31:23 crc kubenswrapper[4750]: I0309 18:31:23.949725 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 09 18:31:24 crc kubenswrapper[4750]: I0309 18:31:24.446071 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 09 18:31:24 crc kubenswrapper[4750]: I0309 18:31:24.455583 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 09 18:31:24 crc kubenswrapper[4750]: I0309 18:31:24.640531 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 09 18:31:24 crc kubenswrapper[4750]: I0309 18:31:24.873857 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 09 18:31:24 crc kubenswrapper[4750]: I0309 18:31:24.898131 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.012872 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.301478 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.482587 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.572837 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.606600 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.802951 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.853017 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.866833 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.881392 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 09 18:31:25 crc kubenswrapper[4750]: I0309 18:31:25.979455 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.040621 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.167703 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.219931 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.595298 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.630913 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.759768 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.765748 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.789649 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.790766 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.856000 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.973012 4750 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 09 18:31:26 crc kubenswrapper[4750]: I0309 18:31:26.980530 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.029470 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.059478 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.085873 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.220674 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.384585 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.569496 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.658449 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.692734 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.736140 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 09 18:31:27 crc kubenswrapper[4750]: I0309 18:31:27.992601 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.043857 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.046784 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.171754 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.211735 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.243155 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.303285 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.341296 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.660016 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.877607 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.892227 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.941644 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.942465 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 09 18:31:28 crc kubenswrapper[4750]: I0309 18:31:28.954053 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.103262 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.120174 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.388406 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.406499 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.408067 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.433697 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.485444 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.517655 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.537098 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.597064 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.675801 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.804613 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.848601 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.865104 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.902468 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.906025 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.908928 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 09 18:31:29 crc kubenswrapper[4750]: I0309 18:31:29.917055 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.049574 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.055322 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.096491 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.149907 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.151456 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.153048 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.222956 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.447449 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.520510 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.678539 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.692094 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.728209 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.871082 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.955319 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.964062 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 09 18:31:30 crc kubenswrapper[4750]: I0309 18:31:30.967301 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.126216 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.192408 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.231204 4750 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.240658 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.302576 4750 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.304154 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=38.304135708 podStartE2EDuration="38.304135708s" podCreationTimestamp="2026-03-09 18:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:31:13.870798836 +0000 UTC m=+355.213271254" watchObservedRunningTime="2026-03-09 18:31:31.304135708 +0000 UTC m=+372.646608106" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.307913 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-jcrcb"] Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.307971 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.312217 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.330051 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.330029071 podStartE2EDuration="18.330029071s" podCreationTimestamp="2026-03-09 18:31:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:31:31.326329377 +0000 UTC m=+372.668801795" watchObservedRunningTime="2026-03-09 18:31:31.330029071 +0000 UTC m=+372.672501469" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.380849 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f821a876-eab1-4298-b06c-9560fb385085" path="/var/lib/kubelet/pods/f821a876-eab1-4298-b06c-9560fb385085/volumes" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.439117 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.445208 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.485092 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.485297 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.581726 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.602185 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.683777 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.811920 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.880859 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.924557 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 09 18:31:31 crc kubenswrapper[4750]: I0309 18:31:31.997082 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.007376 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.147376 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.179242 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.284408 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.308114 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.325840 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.354867 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.356106 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.356504 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.426538 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.478251 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.490065 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.514818 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.567572 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.586705 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.657343 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.679986 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.714170 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.721589 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.725165 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.748436 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.830151 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 09 18:31:32 crc kubenswrapper[4750]: I0309 18:31:32.845489 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.016277 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.068762 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.361229 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.373416 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.417373 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.555073 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.581870 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.642975 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.654289 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.684001 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.704429 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.780159 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.795407 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.865724 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.957438 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.965056 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 18:31:33 crc kubenswrapper[4750]: I0309 18:31:33.996931 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.026130 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.070733 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.150451 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.176248 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.272941 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.473006 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.588121 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.650518 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.783588 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.850197 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.896679 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 09 18:31:34 crc kubenswrapper[4750]: I0309 18:31:34.997999 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.080827 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.089518 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk"] Mar 09 18:31:35 crc kubenswrapper[4750]: E0309 18:31:35.089963 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f821a876-eab1-4298-b06c-9560fb385085" containerName="oauth-openshift" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.089990 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f821a876-eab1-4298-b06c-9560fb385085" containerName="oauth-openshift" Mar 09 18:31:35 crc kubenswrapper[4750]: E0309 18:31:35.090019 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" containerName="installer" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.090028 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" containerName="installer" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.090157 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f821a876-eab1-4298-b06c-9560fb385085" containerName="oauth-openshift" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.090176 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e321a00-cd51-4fc5-b20f-8a6c5d7f3243" containerName="installer" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.090763 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.092596 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.093577 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.096365 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.098734 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.098761 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.098810 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.098857 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.098915 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.099051 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.099086 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.099460 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.099524 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.103009 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.116305 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.117532 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.118847 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk"] Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.131907 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.178756 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190031 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-login\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190128 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-audit-dir\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190184 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190225 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190260 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rj42\" (UniqueName: \"kubernetes.io/projected/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-kube-api-access-8rj42\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190298 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-audit-policies\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190375 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190425 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190460 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-session\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190510 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190617 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190758 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190797 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-error\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.190845 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.292545 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.293209 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-session\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.293364 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.293538 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.293720 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.293863 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-error\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.294019 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.294175 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-login\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.294341 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-audit-dir\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.294513 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.294678 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.294761 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.294511 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-audit-dir\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.295246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.295303 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.295425 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rj42\" (UniqueName: \"kubernetes.io/projected/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-kube-api-access-8rj42\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.295583 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-audit-policies\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.295701 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.297739 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-audit-policies\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.300041 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.301426 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-session\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.301529 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.301548 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.301950 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-error\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.302012 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.302442 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-template-login\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.313079 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.314451 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.319167 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rj42\" (UniqueName: \"kubernetes.io/projected/a7c6dc3d-223a-4fb4-bd6d-d275043173b6-kube-api-access-8rj42\") pod \"oauth-openshift-bbb7d99d8-7tgvk\" (UID: \"a7c6dc3d-223a-4fb4-bd6d-d275043173b6\") " pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.374621 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.390691 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.416365 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.431098 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.481973 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.541673 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.565377 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.658929 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.677698 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.697433 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.713656 4750 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.766282 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.786314 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.884200 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk"] Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.895548 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.934922 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.938767 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.946337 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 09 18:31:35 crc kubenswrapper[4750]: I0309 18:31:35.979747 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.019406 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" event={"ID":"a7c6dc3d-223a-4fb4-bd6d-d275043173b6","Type":"ContainerStarted","Data":"532162a26219f198031e164e404339cb1ef5ec73b7e1ec50448ba74dc0476d83"} Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.161608 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.180001 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.237189 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.242869 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.288806 4750 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.289083 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb" gracePeriod=5 Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.316104 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.413978 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.429210 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.528089 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.674185 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.751757 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.870454 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 09 18:31:36 crc kubenswrapper[4750]: I0309 18:31:36.901941 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.028395 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" event={"ID":"a7c6dc3d-223a-4fb4-bd6d-d275043173b6","Type":"ContainerStarted","Data":"f6417eeb01ed69b9155a9dbf61d2a063b113c6025779c532530cbfc9d1e9ead8"} Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.028567 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.030842 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.037024 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.051744 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-bbb7d99d8-7tgvk" podStartSLOduration=46.051673891 podStartE2EDuration="46.051673891s" podCreationTimestamp="2026-03-09 18:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:31:37.050419756 +0000 UTC m=+378.392892154" watchObservedRunningTime="2026-03-09 18:31:37.051673891 +0000 UTC m=+378.394146309" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.059870 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.174457 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.209088 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.323368 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.329149 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.433743 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.496015 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.701325 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.726782 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.735760 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.746064 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 09 18:31:37 crc kubenswrapper[4750]: I0309 18:31:37.784915 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.139157 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.273526 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.317928 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.388424 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.596437 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.650499 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.657508 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.669733 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.689592 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.727959 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.755688 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.835234 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 18:31:38 crc kubenswrapper[4750]: I0309 18:31:38.842216 4750 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.031575 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.268304 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.282572 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.317758 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.593428 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.747710 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.773656 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 09 18:31:39 crc kubenswrapper[4750]: I0309 18:31:39.858863 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 09 18:31:40 crc kubenswrapper[4750]: I0309 18:31:40.014988 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 09 18:31:40 crc kubenswrapper[4750]: I0309 18:31:40.018419 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 09 18:31:40 crc kubenswrapper[4750]: I0309 18:31:40.035258 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 09 18:31:40 crc kubenswrapper[4750]: I0309 18:31:40.182492 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 18:31:40 crc kubenswrapper[4750]: I0309 18:31:40.309383 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 09 18:31:40 crc kubenswrapper[4750]: I0309 18:31:40.634893 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 09 18:31:41 crc kubenswrapper[4750]: I0309 18:31:41.049470 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 18:31:41 crc kubenswrapper[4750]: I0309 18:31:41.070188 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 09 18:31:41 crc kubenswrapper[4750]: I0309 18:31:41.109854 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 09 18:31:41 crc kubenswrapper[4750]: I0309 18:31:41.190071 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 09 18:31:41 crc kubenswrapper[4750]: I0309 18:31:41.620532 4750 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 09 18:31:41 crc kubenswrapper[4750]: I0309 18:31:41.873008 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 09 18:31:41 crc kubenswrapper[4750]: I0309 18:31:41.873541 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.008902 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.008954 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009021 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009045 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009059 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009163 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009225 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009289 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009308 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009618 4750 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009681 4750 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009699 4750 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.009712 4750 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.017262 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.064185 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.064242 4750 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb" exitCode=137 Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.064353 4750 scope.go:117] "RemoveContainer" containerID="5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.064520 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.088219 4750 scope.go:117] "RemoveContainer" containerID="5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb" Mar 09 18:31:42 crc kubenswrapper[4750]: E0309 18:31:42.088805 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb\": container with ID starting with 5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb not found: ID does not exist" containerID="5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.088901 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb"} err="failed to get container status \"5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb\": rpc error: code = NotFound desc = could not find container \"5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb\": container with ID starting with 5f3e3d7a608a4c5fa17910c976664053d3d26c8f76b27af4b7ff69ce12f1fddb not found: ID does not exist" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.110932 4750 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.145796 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 09 18:31:42 crc kubenswrapper[4750]: I0309 18:31:42.225819 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 09 18:31:43 crc kubenswrapper[4750]: I0309 18:31:43.300741 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 09 18:31:43 crc kubenswrapper[4750]: I0309 18:31:43.381827 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 09 18:31:43 crc kubenswrapper[4750]: I0309 18:31:43.382290 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 09 18:31:43 crc kubenswrapper[4750]: I0309 18:31:43.394245 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 09 18:31:43 crc kubenswrapper[4750]: I0309 18:31:43.394304 4750 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="5a453f67-8fba-4520-a09b-5cdc91bb79ee" Mar 09 18:31:43 crc kubenswrapper[4750]: I0309 18:31:43.404901 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 09 18:31:43 crc kubenswrapper[4750]: I0309 18:31:43.404963 4750 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="5a453f67-8fba-4520-a09b-5cdc91bb79ee" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.171311 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551352-mqsjs"] Mar 09 18:32:00 crc kubenswrapper[4750]: E0309 18:32:00.172205 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.172218 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.172328 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.172778 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551352-mqsjs" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.176558 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.176722 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.176997 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.194034 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8n2n\" (UniqueName: \"kubernetes.io/projected/160b58c3-8419-4f38-8aae-e02628170cb7-kube-api-access-j8n2n\") pod \"auto-csr-approver-29551352-mqsjs\" (UID: \"160b58c3-8419-4f38-8aae-e02628170cb7\") " pod="openshift-infra/auto-csr-approver-29551352-mqsjs" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.213426 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551352-mqsjs"] Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.295564 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8n2n\" (UniqueName: \"kubernetes.io/projected/160b58c3-8419-4f38-8aae-e02628170cb7-kube-api-access-j8n2n\") pod \"auto-csr-approver-29551352-mqsjs\" (UID: \"160b58c3-8419-4f38-8aae-e02628170cb7\") " pod="openshift-infra/auto-csr-approver-29551352-mqsjs" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.325378 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8n2n\" (UniqueName: \"kubernetes.io/projected/160b58c3-8419-4f38-8aae-e02628170cb7-kube-api-access-j8n2n\") pod \"auto-csr-approver-29551352-mqsjs\" (UID: \"160b58c3-8419-4f38-8aae-e02628170cb7\") " pod="openshift-infra/auto-csr-approver-29551352-mqsjs" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.491064 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551352-mqsjs" Mar 09 18:32:00 crc kubenswrapper[4750]: I0309 18:32:00.888562 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551352-mqsjs"] Mar 09 18:32:01 crc kubenswrapper[4750]: I0309 18:32:01.210286 4750 generic.go:334] "Generic (PLEG): container finished" podID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerID="0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8" exitCode=0 Mar 09 18:32:01 crc kubenswrapper[4750]: I0309 18:32:01.210386 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" event={"ID":"39ea59b4-900a-43ce-9cf5-b5d029ff0ada","Type":"ContainerDied","Data":"0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8"} Mar 09 18:32:01 crc kubenswrapper[4750]: I0309 18:32:01.211023 4750 scope.go:117] "RemoveContainer" containerID="0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8" Mar 09 18:32:01 crc kubenswrapper[4750]: I0309 18:32:01.214200 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551352-mqsjs" event={"ID":"160b58c3-8419-4f38-8aae-e02628170cb7","Type":"ContainerStarted","Data":"29425ced2441772ad5d07f5057e79b248532eaa8b18f7983d3654600ea68d96e"} Mar 09 18:32:02 crc kubenswrapper[4750]: I0309 18:32:02.225126 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" event={"ID":"39ea59b4-900a-43ce-9cf5-b5d029ff0ada","Type":"ContainerStarted","Data":"b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8"} Mar 09 18:32:02 crc kubenswrapper[4750]: I0309 18:32:02.226049 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:32:02 crc kubenswrapper[4750]: I0309 18:32:02.230477 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:32:03 crc kubenswrapper[4750]: I0309 18:32:03.234509 4750 generic.go:334] "Generic (PLEG): container finished" podID="160b58c3-8419-4f38-8aae-e02628170cb7" containerID="7cd34d7145390006e08fbc6ead05527af5e8be76b668219103e4202e44c5a96c" exitCode=0 Mar 09 18:32:03 crc kubenswrapper[4750]: I0309 18:32:03.234723 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551352-mqsjs" event={"ID":"160b58c3-8419-4f38-8aae-e02628170cb7","Type":"ContainerDied","Data":"7cd34d7145390006e08fbc6ead05527af5e8be76b668219103e4202e44c5a96c"} Mar 09 18:32:04 crc kubenswrapper[4750]: I0309 18:32:04.500107 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551352-mqsjs" Mar 09 18:32:04 crc kubenswrapper[4750]: I0309 18:32:04.656597 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8n2n\" (UniqueName: \"kubernetes.io/projected/160b58c3-8419-4f38-8aae-e02628170cb7-kube-api-access-j8n2n\") pod \"160b58c3-8419-4f38-8aae-e02628170cb7\" (UID: \"160b58c3-8419-4f38-8aae-e02628170cb7\") " Mar 09 18:32:04 crc kubenswrapper[4750]: I0309 18:32:04.665088 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/160b58c3-8419-4f38-8aae-e02628170cb7-kube-api-access-j8n2n" (OuterVolumeSpecName: "kube-api-access-j8n2n") pod "160b58c3-8419-4f38-8aae-e02628170cb7" (UID: "160b58c3-8419-4f38-8aae-e02628170cb7"). InnerVolumeSpecName "kube-api-access-j8n2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:32:04 crc kubenswrapper[4750]: I0309 18:32:04.758404 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8n2n\" (UniqueName: \"kubernetes.io/projected/160b58c3-8419-4f38-8aae-e02628170cb7-kube-api-access-j8n2n\") on node \"crc\" DevicePath \"\"" Mar 09 18:32:05 crc kubenswrapper[4750]: I0309 18:32:05.248803 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551352-mqsjs" event={"ID":"160b58c3-8419-4f38-8aae-e02628170cb7","Type":"ContainerDied","Data":"29425ced2441772ad5d07f5057e79b248532eaa8b18f7983d3654600ea68d96e"} Mar 09 18:32:05 crc kubenswrapper[4750]: I0309 18:32:05.248876 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29425ced2441772ad5d07f5057e79b248532eaa8b18f7983d3654600ea68d96e" Mar 09 18:32:05 crc kubenswrapper[4750]: I0309 18:32:05.248938 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551352-mqsjs" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.864925 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nhclf"] Mar 09 18:32:44 crc kubenswrapper[4750]: E0309 18:32:44.866116 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="160b58c3-8419-4f38-8aae-e02628170cb7" containerName="oc" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.866136 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="160b58c3-8419-4f38-8aae-e02628170cb7" containerName="oc" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.866275 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="160b58c3-8419-4f38-8aae-e02628170cb7" containerName="oc" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.866848 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.879765 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nhclf"] Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977094 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-bound-sa-token\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977171 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90117dea-9764-43d2-8275-e0419630cd7c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977215 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90117dea-9764-43d2-8275-e0419630cd7c-registry-certificates\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977255 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977354 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90117dea-9764-43d2-8275-e0419630cd7c-trusted-ca\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977393 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90117dea-9764-43d2-8275-e0419630cd7c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977432 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-registry-tls\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:44 crc kubenswrapper[4750]: I0309 18:32:44.977464 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft9pm\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-kube-api-access-ft9pm\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.006300 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.079463 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90117dea-9764-43d2-8275-e0419630cd7c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.079525 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90117dea-9764-43d2-8275-e0419630cd7c-registry-certificates\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.079565 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90117dea-9764-43d2-8275-e0419630cd7c-trusted-ca\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.079599 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90117dea-9764-43d2-8275-e0419630cd7c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.079634 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-registry-tls\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.079679 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft9pm\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-kube-api-access-ft9pm\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.079710 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-bound-sa-token\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.080847 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90117dea-9764-43d2-8275-e0419630cd7c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.081757 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90117dea-9764-43d2-8275-e0419630cd7c-trusted-ca\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.083511 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90117dea-9764-43d2-8275-e0419630cd7c-registry-certificates\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.090062 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90117dea-9764-43d2-8275-e0419630cd7c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.097540 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-registry-tls\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.100071 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft9pm\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-kube-api-access-ft9pm\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.101322 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90117dea-9764-43d2-8275-e0419630cd7c-bound-sa-token\") pod \"image-registry-66df7c8f76-nhclf\" (UID: \"90117dea-9764-43d2-8275-e0419630cd7c\") " pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.189976 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:45 crc kubenswrapper[4750]: I0309 18:32:45.628756 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nhclf"] Mar 09 18:32:46 crc kubenswrapper[4750]: I0309 18:32:46.559712 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" event={"ID":"90117dea-9764-43d2-8275-e0419630cd7c","Type":"ContainerStarted","Data":"69b9cb85eb11c918c98f5f63d1fcc5a43f197ba155c0b1af24b68d5cdf15fe22"} Mar 09 18:32:46 crc kubenswrapper[4750]: I0309 18:32:46.559787 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" event={"ID":"90117dea-9764-43d2-8275-e0419630cd7c","Type":"ContainerStarted","Data":"80ef3a387280ebe56a760734b526d964cf4ff3e30f2f98490667b4f5b96761f5"} Mar 09 18:32:46 crc kubenswrapper[4750]: I0309 18:32:46.559912 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:32:46 crc kubenswrapper[4750]: I0309 18:32:46.586372 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" podStartSLOduration=2.5863551190000003 podStartE2EDuration="2.586355119s" podCreationTimestamp="2026-03-09 18:32:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:32:46.582999066 +0000 UTC m=+447.925471504" watchObservedRunningTime="2026-03-09 18:32:46.586355119 +0000 UTC m=+447.928827517" Mar 09 18:32:51 crc kubenswrapper[4750]: I0309 18:32:51.744154 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:32:51 crc kubenswrapper[4750]: I0309 18:32:51.744909 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:33:05 crc kubenswrapper[4750]: I0309 18:33:05.199906 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nhclf" Mar 09 18:33:05 crc kubenswrapper[4750]: I0309 18:33:05.274129 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qmbbg"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.639050 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvzmg"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.640433 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zvzmg" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="registry-server" containerID="cri-o://f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" gracePeriod=30 Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.643615 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rj4zw"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.643941 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rj4zw" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="registry-server" containerID="cri-o://927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" gracePeriod=30 Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.654703 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c9j6s"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.655448 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" containerID="cri-o://b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8" gracePeriod=30 Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.680305 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk4f9"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.680670 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jk4f9" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="registry-server" containerID="cri-o://552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a" gracePeriod=30 Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.687345 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxwd2"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.687609 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lxwd2" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="registry-server" containerID="cri-o://580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e" gracePeriod=30 Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.696609 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2nkw4"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.697424 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.703714 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2nkw4"] Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.773523 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.773601 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kxnv\" (UniqueName: \"kubernetes.io/projected/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-kube-api-access-2kxnv\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.773675 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.875300 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.875395 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kxnv\" (UniqueName: \"kubernetes.io/projected/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-kube-api-access-2kxnv\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.875437 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.876982 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.883550 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:20 crc kubenswrapper[4750]: I0309 18:33:20.902489 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kxnv\" (UniqueName: \"kubernetes.io/projected/ab532dd6-c38a-4218-bdb3-b4492f1d5f2f-kube-api-access-2kxnv\") pod \"marketplace-operator-79b997595-2nkw4\" (UID: \"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f\") " pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:21 crc kubenswrapper[4750]: I0309 18:33:21.073986 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:21 crc kubenswrapper[4750]: I0309 18:33:21.745965 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:33:21 crc kubenswrapper[4750]: I0309 18:33:21.746048 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.818506 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7 is running failed: container process not found" containerID="927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.818878 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67 is running failed: container process not found" containerID="f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.824496 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67 is running failed: container process not found" containerID="f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.824584 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7 is running failed: container process not found" containerID="927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.825899 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7 is running failed: container process not found" containerID="927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.825946 4750 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-rj4zw" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="registry-server" Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.832563 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67 is running failed: container process not found" containerID="f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 18:33:21 crc kubenswrapper[4750]: E0309 18:33:21.832703 4750 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-zvzmg" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="registry-server" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.197558 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.246559 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.256308 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.274473 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.295756 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-878gv\" (UniqueName: \"kubernetes.io/projected/41608683-d508-4374-a570-0e93ca339ffb-kube-api-access-878gv\") pod \"41608683-d508-4374-a570-0e93ca339ffb\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.295849 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-utilities\") pod \"41608683-d508-4374-a570-0e93ca339ffb\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.295952 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-catalog-content\") pod \"41608683-d508-4374-a570-0e93ca339ffb\" (UID: \"41608683-d508-4374-a570-0e93ca339ffb\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.298731 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-utilities" (OuterVolumeSpecName: "utilities") pod "41608683-d508-4374-a570-0e93ca339ffb" (UID: "41608683-d508-4374-a570-0e93ca339ffb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.309112 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41608683-d508-4374-a570-0e93ca339ffb-kube-api-access-878gv" (OuterVolumeSpecName: "kube-api-access-878gv") pod "41608683-d508-4374-a570-0e93ca339ffb" (UID: "41608683-d508-4374-a570-0e93ca339ffb"). InnerVolumeSpecName "kube-api-access-878gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397345 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbbbz\" (UniqueName: \"kubernetes.io/projected/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-kube-api-access-cbbbz\") pod \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397461 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-operator-metrics\") pod \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397485 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-trusted-ca\") pod \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\" (UID: \"39ea59b4-900a-43ce-9cf5-b5d029ff0ada\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397520 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btsjs\" (UniqueName: \"kubernetes.io/projected/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-kube-api-access-btsjs\") pod \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397570 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-catalog-content\") pod \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397595 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-utilities\") pod \"2de1257e-0c9d-486f-92da-2e0525941f7e\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397623 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-utilities\") pod \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\" (UID: \"ab00ccba-a824-42a2-a9fc-43139bcf9a0e\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397666 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-catalog-content\") pod \"2de1257e-0c9d-486f-92da-2e0525941f7e\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.397701 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x27vz\" (UniqueName: \"kubernetes.io/projected/2de1257e-0c9d-486f-92da-2e0525941f7e-kube-api-access-x27vz\") pod \"2de1257e-0c9d-486f-92da-2e0525941f7e\" (UID: \"2de1257e-0c9d-486f-92da-2e0525941f7e\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.398029 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-878gv\" (UniqueName: \"kubernetes.io/projected/41608683-d508-4374-a570-0e93ca339ffb-kube-api-access-878gv\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.398050 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.403287 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "39ea59b4-900a-43ce-9cf5-b5d029ff0ada" (UID: "39ea59b4-900a-43ce-9cf5-b5d029ff0ada"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.404190 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-utilities" (OuterVolumeSpecName: "utilities") pod "ab00ccba-a824-42a2-a9fc-43139bcf9a0e" (UID: "ab00ccba-a824-42a2-a9fc-43139bcf9a0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.410881 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "39ea59b4-900a-43ce-9cf5-b5d029ff0ada" (UID: "39ea59b4-900a-43ce-9cf5-b5d029ff0ada"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.414855 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-kube-api-access-btsjs" (OuterVolumeSpecName: "kube-api-access-btsjs") pod "ab00ccba-a824-42a2-a9fc-43139bcf9a0e" (UID: "ab00ccba-a824-42a2-a9fc-43139bcf9a0e"). InnerVolumeSpecName "kube-api-access-btsjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.416712 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-kube-api-access-cbbbz" (OuterVolumeSpecName: "kube-api-access-cbbbz") pod "39ea59b4-900a-43ce-9cf5-b5d029ff0ada" (UID: "39ea59b4-900a-43ce-9cf5-b5d029ff0ada"). InnerVolumeSpecName "kube-api-access-cbbbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.429807 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-utilities" (OuterVolumeSpecName: "utilities") pod "2de1257e-0c9d-486f-92da-2e0525941f7e" (UID: "2de1257e-0c9d-486f-92da-2e0525941f7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.436126 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de1257e-0c9d-486f-92da-2e0525941f7e-kube-api-access-x27vz" (OuterVolumeSpecName: "kube-api-access-x27vz") pod "2de1257e-0c9d-486f-92da-2e0525941f7e" (UID: "2de1257e-0c9d-486f-92da-2e0525941f7e"). InnerVolumeSpecName "kube-api-access-x27vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.473930 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2nkw4"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.496981 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2de1257e-0c9d-486f-92da-2e0525941f7e" (UID: "2de1257e-0c9d-486f-92da-2e0525941f7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.500728 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41608683-d508-4374-a570-0e93ca339ffb" (UID: "41608683-d508-4374-a570-0e93ca339ffb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.514962 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515008 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515022 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de1257e-0c9d-486f-92da-2e0525941f7e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515039 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x27vz\" (UniqueName: \"kubernetes.io/projected/2de1257e-0c9d-486f-92da-2e0525941f7e-kube-api-access-x27vz\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515054 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbbbz\" (UniqueName: \"kubernetes.io/projected/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-kube-api-access-cbbbz\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515066 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41608683-d508-4374-a570-0e93ca339ffb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515080 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515093 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39ea59b4-900a-43ce-9cf5-b5d029ff0ada-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.515105 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btsjs\" (UniqueName: \"kubernetes.io/projected/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-kube-api-access-btsjs\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.528463 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab00ccba-a824-42a2-a9fc-43139bcf9a0e" (UID: "ab00ccba-a824-42a2-a9fc-43139bcf9a0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.616793 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.620689 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab00ccba-a824-42a2-a9fc-43139bcf9a0e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.721987 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-catalog-content\") pod \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.722142 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw7mj\" (UniqueName: \"kubernetes.io/projected/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-kube-api-access-lw7mj\") pod \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.722172 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-utilities\") pod \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\" (UID: \"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082\") " Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.723320 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-utilities" (OuterVolumeSpecName: "utilities") pod "8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" (UID: "8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.729312 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-kube-api-access-lw7mj" (OuterVolumeSpecName: "kube-api-access-lw7mj") pod "8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" (UID: "8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082"). InnerVolumeSpecName "kube-api-access-lw7mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.824235 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw7mj\" (UniqueName: \"kubernetes.io/projected/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-kube-api-access-lw7mj\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.826089 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.833938 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" event={"ID":"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f","Type":"ContainerStarted","Data":"4ab6111db3751e0960fc418682738432f535ac502fffd3eab15a1aaba4d2592e"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.834008 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" event={"ID":"ab532dd6-c38a-4218-bdb3-b4492f1d5f2f","Type":"ContainerStarted","Data":"a7ce13054e3a5af57effb5b2c8af912e32e497af7997aed2c38e80307a13c334"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.834220 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.835666 4750 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2nkw4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.835726 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" podUID="ab532dd6-c38a-4218-bdb3-b4492f1d5f2f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.838453 4750 generic.go:334] "Generic (PLEG): container finished" podID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerID="552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a" exitCode=0 Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.838562 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk4f9" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.838618 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk4f9" event={"ID":"2de1257e-0c9d-486f-92da-2e0525941f7e","Type":"ContainerDied","Data":"552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.838680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk4f9" event={"ID":"2de1257e-0c9d-486f-92da-2e0525941f7e","Type":"ContainerDied","Data":"7505bda44a51b5cb2d2d01982524f8c840a7bf927107e2587fb6aed9052a5ee2"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.838710 4750 scope.go:117] "RemoveContainer" containerID="552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.845344 4750 generic.go:334] "Generic (PLEG): container finished" podID="41608683-d508-4374-a570-0e93ca339ffb" containerID="927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" exitCode=0 Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.845534 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rj4zw" event={"ID":"41608683-d508-4374-a570-0e93ca339ffb","Type":"ContainerDied","Data":"927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.845677 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rj4zw" event={"ID":"41608683-d508-4374-a570-0e93ca339ffb","Type":"ContainerDied","Data":"2b22137fe94cb982150ca33e9853cc68d6d5b78e342e9d5c86cdac63751dbcf3"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.845569 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rj4zw" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.849311 4750 generic.go:334] "Generic (PLEG): container finished" podID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerID="580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e" exitCode=0 Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.849412 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxwd2" event={"ID":"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082","Type":"ContainerDied","Data":"580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.849465 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxwd2" event={"ID":"8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082","Type":"ContainerDied","Data":"9af23198bc0b1f7e1d2c46fa1368dcc4c78a418be0fe63e0fc2db838c92b33fe"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.849481 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxwd2" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.855297 4750 generic.go:334] "Generic (PLEG): container finished" podID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerID="b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8" exitCode=0 Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.855423 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" event={"ID":"39ea59b4-900a-43ce-9cf5-b5d029ff0ada","Type":"ContainerDied","Data":"b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.855465 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" event={"ID":"39ea59b4-900a-43ce-9cf5-b5d029ff0ada","Type":"ContainerDied","Data":"bacaa8f1a6265292ac23fa9afb204b8509f66bcf22d8e01c8e5a17678b7db826"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.855555 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c9j6s" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.859445 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" podStartSLOduration=2.859418897 podStartE2EDuration="2.859418897s" podCreationTimestamp="2026-03-09 18:33:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:33:22.85092648 +0000 UTC m=+484.193398888" watchObservedRunningTime="2026-03-09 18:33:22.859418897 +0000 UTC m=+484.201891295" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.859857 4750 scope.go:117] "RemoveContainer" containerID="425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.861019 4750 generic.go:334] "Generic (PLEG): container finished" podID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerID="f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" exitCode=0 Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.861077 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvzmg" event={"ID":"ab00ccba-a824-42a2-a9fc-43139bcf9a0e","Type":"ContainerDied","Data":"f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.861116 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvzmg" event={"ID":"ab00ccba-a824-42a2-a9fc-43139bcf9a0e","Type":"ContainerDied","Data":"da7388fecb1b148c9444914252ca05d16e962fa136b3f760811b5a5bfe63db3d"} Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.861269 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvzmg" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.867271 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" (UID: "8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.878717 4750 scope.go:117] "RemoveContainer" containerID="211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.888971 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rj4zw"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.920012 4750 scope.go:117] "RemoveContainer" containerID="552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a" Mar 09 18:33:22 crc kubenswrapper[4750]: E0309 18:33:22.920663 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a\": container with ID starting with 552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a not found: ID does not exist" containerID="552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.920700 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a"} err="failed to get container status \"552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a\": rpc error: code = NotFound desc = could not find container \"552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a\": container with ID starting with 552b28d58fee1b353526d4d2fbaec7457ed7da00c6326dcd3c8a018770e8670a not found: ID does not exist" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.920725 4750 scope.go:117] "RemoveContainer" containerID="425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b" Mar 09 18:33:22 crc kubenswrapper[4750]: E0309 18:33:22.921236 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b\": container with ID starting with 425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b not found: ID does not exist" containerID="425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.921294 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b"} err="failed to get container status \"425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b\": rpc error: code = NotFound desc = could not find container \"425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b\": container with ID starting with 425356d2c107c50c7435b44bf830f3c16f93323d34b398da02865c9d82d23a7b not found: ID does not exist" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.921334 4750 scope.go:117] "RemoveContainer" containerID="211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0" Mar 09 18:33:22 crc kubenswrapper[4750]: E0309 18:33:22.921622 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0\": container with ID starting with 211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0 not found: ID does not exist" containerID="211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.921659 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0"} err="failed to get container status \"211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0\": rpc error: code = NotFound desc = could not find container \"211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0\": container with ID starting with 211184979d65c1bcabc84b0f16b083f7a483b8898d569c6a1e9caa80ab0998d0 not found: ID does not exist" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.921678 4750 scope.go:117] "RemoveContainer" containerID="927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.925380 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rj4zw"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.928691 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.933882 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk4f9"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.941872 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk4f9"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.948041 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c9j6s"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.949189 4750 scope.go:117] "RemoveContainer" containerID="8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.959293 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c9j6s"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.966929 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvzmg"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.970298 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zvzmg"] Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.976082 4750 scope.go:117] "RemoveContainer" containerID="fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.990854 4750 scope.go:117] "RemoveContainer" containerID="927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" Mar 09 18:33:22 crc kubenswrapper[4750]: E0309 18:33:22.991225 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7\": container with ID starting with 927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7 not found: ID does not exist" containerID="927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.991267 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7"} err="failed to get container status \"927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7\": rpc error: code = NotFound desc = could not find container \"927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7\": container with ID starting with 927fcc6bdcfecc63896945a32a479aef85412a226a78084d5d9e3d1e4c58e5c7 not found: ID does not exist" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.991319 4750 scope.go:117] "RemoveContainer" containerID="8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585" Mar 09 18:33:22 crc kubenswrapper[4750]: E0309 18:33:22.991538 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585\": container with ID starting with 8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585 not found: ID does not exist" containerID="8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.991563 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585"} err="failed to get container status \"8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585\": rpc error: code = NotFound desc = could not find container \"8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585\": container with ID starting with 8e2fb6615e676df7f5996504a085146c512d31d610a55d7e3b60ba22d2ebc585 not found: ID does not exist" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.991581 4750 scope.go:117] "RemoveContainer" containerID="fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7" Mar 09 18:33:22 crc kubenswrapper[4750]: E0309 18:33:22.991785 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7\": container with ID starting with fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7 not found: ID does not exist" containerID="fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.991808 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7"} err="failed to get container status \"fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7\": rpc error: code = NotFound desc = could not find container \"fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7\": container with ID starting with fe2b5b87bbd90d49660c491fe13ee7a27077e019e9b4c850d56f78548d5e24c7 not found: ID does not exist" Mar 09 18:33:22 crc kubenswrapper[4750]: I0309 18:33:22.991821 4750 scope.go:117] "RemoveContainer" containerID="580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.006774 4750 scope.go:117] "RemoveContainer" containerID="8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.030720 4750 scope.go:117] "RemoveContainer" containerID="eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.052008 4750 scope.go:117] "RemoveContainer" containerID="580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.052536 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e\": container with ID starting with 580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e not found: ID does not exist" containerID="580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.052576 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e"} err="failed to get container status \"580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e\": rpc error: code = NotFound desc = could not find container \"580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e\": container with ID starting with 580e8dee77e42b15934948ef3f87401fe1815f943b0c88d4b1040c24332c972e not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.052608 4750 scope.go:117] "RemoveContainer" containerID="8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.052946 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c\": container with ID starting with 8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c not found: ID does not exist" containerID="8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.052967 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c"} err="failed to get container status \"8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c\": rpc error: code = NotFound desc = could not find container \"8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c\": container with ID starting with 8380a695dd1c11137b333ff090911dd2d0fb3465935a3211da0ff7654236d91c not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.052982 4750 scope.go:117] "RemoveContainer" containerID="eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.053225 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092\": container with ID starting with eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092 not found: ID does not exist" containerID="eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.053250 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092"} err="failed to get container status \"eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092\": rpc error: code = NotFound desc = could not find container \"eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092\": container with ID starting with eafa7e61946611d82c1979fa5e24647a2dd1d749598f36ae3c0bf950d7222092 not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.053265 4750 scope.go:117] "RemoveContainer" containerID="b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.069692 4750 scope.go:117] "RemoveContainer" containerID="0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.084477 4750 scope.go:117] "RemoveContainer" containerID="b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.085081 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8\": container with ID starting with b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8 not found: ID does not exist" containerID="b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.085112 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8"} err="failed to get container status \"b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8\": rpc error: code = NotFound desc = could not find container \"b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8\": container with ID starting with b6afea4ab03abbc82a62f946e3fe557232d515bb71f53ac8b8f18c80ec27ada8 not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.085147 4750 scope.go:117] "RemoveContainer" containerID="0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.085565 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8\": container with ID starting with 0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8 not found: ID does not exist" containerID="0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.085585 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8"} err="failed to get container status \"0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8\": rpc error: code = NotFound desc = could not find container \"0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8\": container with ID starting with 0b43c034ea19068191d86cce90937b9e9271c163b48d51aa12b54fd1a08b1fb8 not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.085598 4750 scope.go:117] "RemoveContainer" containerID="f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.099436 4750 scope.go:117] "RemoveContainer" containerID="97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.115326 4750 scope.go:117] "RemoveContainer" containerID="0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.130316 4750 scope.go:117] "RemoveContainer" containerID="f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.131520 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67\": container with ID starting with f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67 not found: ID does not exist" containerID="f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.131557 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67"} err="failed to get container status \"f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67\": rpc error: code = NotFound desc = could not find container \"f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67\": container with ID starting with f555248f75c8c053a55caef48a61b70ae8671846ab3e1a3595b98623f0030d67 not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.131583 4750 scope.go:117] "RemoveContainer" containerID="97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.132077 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426\": container with ID starting with 97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426 not found: ID does not exist" containerID="97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.132102 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426"} err="failed to get container status \"97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426\": rpc error: code = NotFound desc = could not find container \"97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426\": container with ID starting with 97444a629a12abec2cff8f129678b9398ee310fc17d336146d0cdd1a029c4426 not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.132128 4750 scope.go:117] "RemoveContainer" containerID="0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11" Mar 09 18:33:23 crc kubenswrapper[4750]: E0309 18:33:23.132383 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11\": container with ID starting with 0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11 not found: ID does not exist" containerID="0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.132408 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11"} err="failed to get container status \"0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11\": rpc error: code = NotFound desc = could not find container \"0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11\": container with ID starting with 0e656316a3b7b587a549fa017db207ab7abb971d3646d2ae2bb34e7480b36d11 not found: ID does not exist" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.182591 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxwd2"] Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.189974 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lxwd2"] Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.381220 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" path="/var/lib/kubelet/pods/2de1257e-0c9d-486f-92da-2e0525941f7e/volumes" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.381999 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" path="/var/lib/kubelet/pods/39ea59b4-900a-43ce-9cf5-b5d029ff0ada/volumes" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.382497 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41608683-d508-4374-a570-0e93ca339ffb" path="/var/lib/kubelet/pods/41608683-d508-4374-a570-0e93ca339ffb/volumes" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.383828 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" path="/var/lib/kubelet/pods/8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082/volumes" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.384541 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" path="/var/lib/kubelet/pods/ab00ccba-a824-42a2-a9fc-43139bcf9a0e/volumes" Mar 09 18:33:23 crc kubenswrapper[4750]: I0309 18:33:23.876009 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2nkw4" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850596 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wtbl2"] Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850876 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850889 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850901 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850907 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850915 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850922 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850931 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850937 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850947 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850954 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850962 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850968 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850974 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850980 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.850991 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.850997 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.851007 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851013 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.851023 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851030 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.851039 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851045 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="extract-utilities" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.851056 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851064 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="extract-content" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.851072 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851079 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851184 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab00ccba-a824-42a2-a9fc-43139bcf9a0e" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851202 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851210 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851217 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de1257e-0c9d-486f-92da-2e0525941f7e" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851226 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc3cd0f-236e-4479-9f8e-2f3e1c6fb082" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851238 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="41608683-d508-4374-a570-0e93ca339ffb" containerName="registry-server" Mar 09 18:33:24 crc kubenswrapper[4750]: E0309 18:33:24.851327 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.851335 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ea59b4-900a-43ce-9cf5-b5d029ff0ada" containerName="marketplace-operator" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.852043 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.859958 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.866844 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wtbl2"] Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.974470 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-utilities\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.974548 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-catalog-content\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:24 crc kubenswrapper[4750]: I0309 18:33:24.974617 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt87m\" (UniqueName: \"kubernetes.io/projected/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-kube-api-access-qt87m\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.050008 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-55x9g"] Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.051097 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.055884 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.066752 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-55x9g"] Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.076489 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-utilities\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.076590 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-catalog-content\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.076654 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt87m\" (UniqueName: \"kubernetes.io/projected/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-kube-api-access-qt87m\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.077226 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-catalog-content\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.077419 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-utilities\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.096164 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt87m\" (UniqueName: \"kubernetes.io/projected/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-kube-api-access-qt87m\") pod \"certified-operators-wtbl2\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.177880 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2da9cec-0767-4af5-a433-c9a50eeca1fc-catalog-content\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.178342 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.178400 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vldtw\" (UniqueName: \"kubernetes.io/projected/a2da9cec-0767-4af5-a433-c9a50eeca1fc-kube-api-access-vldtw\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.178472 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2da9cec-0767-4af5-a433-c9a50eeca1fc-utilities\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.279603 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2da9cec-0767-4af5-a433-c9a50eeca1fc-catalog-content\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.279684 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vldtw\" (UniqueName: \"kubernetes.io/projected/a2da9cec-0767-4af5-a433-c9a50eeca1fc-kube-api-access-vldtw\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.279756 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2da9cec-0767-4af5-a433-c9a50eeca1fc-utilities\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.280304 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2da9cec-0767-4af5-a433-c9a50eeca1fc-catalog-content\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.280339 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2da9cec-0767-4af5-a433-c9a50eeca1fc-utilities\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.298090 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vldtw\" (UniqueName: \"kubernetes.io/projected/a2da9cec-0767-4af5-a433-c9a50eeca1fc-kube-api-access-vldtw\") pod \"community-operators-55x9g\" (UID: \"a2da9cec-0767-4af5-a433-c9a50eeca1fc\") " pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.366437 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.407441 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wtbl2"] Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.594177 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-55x9g"] Mar 09 18:33:25 crc kubenswrapper[4750]: W0309 18:33:25.664984 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2da9cec_0767_4af5_a433_c9a50eeca1fc.slice/crio-2359687b2deb80b1b8a22ec254d02913693d843da619e2dd4e607f28ff55c830 WatchSource:0}: Error finding container 2359687b2deb80b1b8a22ec254d02913693d843da619e2dd4e607f28ff55c830: Status 404 returned error can't find the container with id 2359687b2deb80b1b8a22ec254d02913693d843da619e2dd4e607f28ff55c830 Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.898743 4750 generic.go:334] "Generic (PLEG): container finished" podID="a2da9cec-0767-4af5-a433-c9a50eeca1fc" containerID="23a4876bbfcc8c262809d2a4846bd9f9b2b2a00bd91c55c83d112eae4b56035b" exitCode=0 Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.898824 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-55x9g" event={"ID":"a2da9cec-0767-4af5-a433-c9a50eeca1fc","Type":"ContainerDied","Data":"23a4876bbfcc8c262809d2a4846bd9f9b2b2a00bd91c55c83d112eae4b56035b"} Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.898904 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-55x9g" event={"ID":"a2da9cec-0767-4af5-a433-c9a50eeca1fc","Type":"ContainerStarted","Data":"2359687b2deb80b1b8a22ec254d02913693d843da619e2dd4e607f28ff55c830"} Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.902443 4750 generic.go:334] "Generic (PLEG): container finished" podID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerID="31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c" exitCode=0 Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.902676 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtbl2" event={"ID":"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83","Type":"ContainerDied","Data":"31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c"} Mar 09 18:33:25 crc kubenswrapper[4750]: I0309 18:33:25.902951 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtbl2" event={"ID":"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83","Type":"ContainerStarted","Data":"01bd8be7c16c3a58b7410c558ad69a65fd9a10329690ce54b8c7ac02de4e4837"} Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.251472 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-66h4d"] Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.253312 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.257239 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.267307 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-66h4d"] Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.411850 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06088b9b-c876-4cf1-978e-3aae41970e58-catalog-content\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.411947 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06088b9b-c876-4cf1-978e-3aae41970e58-utilities\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.412091 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5fln\" (UniqueName: \"kubernetes.io/projected/06088b9b-c876-4cf1-978e-3aae41970e58-kube-api-access-v5fln\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.451345 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qsrwr"] Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.452745 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.457251 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.463069 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qsrwr"] Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.513014 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06088b9b-c876-4cf1-978e-3aae41970e58-catalog-content\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.513115 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06088b9b-c876-4cf1-978e-3aae41970e58-utilities\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.513157 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5fln\" (UniqueName: \"kubernetes.io/projected/06088b9b-c876-4cf1-978e-3aae41970e58-kube-api-access-v5fln\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.513868 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06088b9b-c876-4cf1-978e-3aae41970e58-utilities\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.513944 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06088b9b-c876-4cf1-978e-3aae41970e58-catalog-content\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.540237 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5fln\" (UniqueName: \"kubernetes.io/projected/06088b9b-c876-4cf1-978e-3aae41970e58-kube-api-access-v5fln\") pod \"redhat-marketplace-66h4d\" (UID: \"06088b9b-c876-4cf1-978e-3aae41970e58\") " pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.615863 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mtrg\" (UniqueName: \"kubernetes.io/projected/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-kube-api-access-6mtrg\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.616060 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-utilities\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.616117 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-catalog-content\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.650797 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.717427 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-utilities\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.717855 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-catalog-content\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.717934 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mtrg\" (UniqueName: \"kubernetes.io/projected/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-kube-api-access-6mtrg\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.718261 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-utilities\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.718480 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-catalog-content\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.740979 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mtrg\" (UniqueName: \"kubernetes.io/projected/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-kube-api-access-6mtrg\") pod \"redhat-operators-qsrwr\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.812363 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.923983 4750 generic.go:334] "Generic (PLEG): container finished" podID="a2da9cec-0767-4af5-a433-c9a50eeca1fc" containerID="1adbc61833df637267dcf6e04df230e983508e9e9daad370a59801010aaa0653" exitCode=0 Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.924065 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-55x9g" event={"ID":"a2da9cec-0767-4af5-a433-c9a50eeca1fc","Type":"ContainerDied","Data":"1adbc61833df637267dcf6e04df230e983508e9e9daad370a59801010aaa0653"} Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.927040 4750 generic.go:334] "Generic (PLEG): container finished" podID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerID="66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834" exitCode=0 Mar 09 18:33:27 crc kubenswrapper[4750]: I0309 18:33:27.927086 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtbl2" event={"ID":"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83","Type":"ContainerDied","Data":"66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834"} Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.026699 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qsrwr"] Mar 09 18:33:28 crc kubenswrapper[4750]: W0309 18:33:28.034944 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdd878ef_1b79_49cd_a6bb_b756449fc7d8.slice/crio-002bb66edcf6f6d7dddf4733fd8e55e8611d2cf193cb2ebb587b044fe496d820 WatchSource:0}: Error finding container 002bb66edcf6f6d7dddf4733fd8e55e8611d2cf193cb2ebb587b044fe496d820: Status 404 returned error can't find the container with id 002bb66edcf6f6d7dddf4733fd8e55e8611d2cf193cb2ebb587b044fe496d820 Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.085553 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-66h4d"] Mar 09 18:33:28 crc kubenswrapper[4750]: W0309 18:33:28.089179 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06088b9b_c876_4cf1_978e_3aae41970e58.slice/crio-fa8af6d2b00757cf9d56715ef86694b04cc725725cf5714842a7eefe95443f4b WatchSource:0}: Error finding container fa8af6d2b00757cf9d56715ef86694b04cc725725cf5714842a7eefe95443f4b: Status 404 returned error can't find the container with id fa8af6d2b00757cf9d56715ef86694b04cc725725cf5714842a7eefe95443f4b Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.936585 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtbl2" event={"ID":"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83","Type":"ContainerStarted","Data":"0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0"} Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.939388 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-55x9g" event={"ID":"a2da9cec-0767-4af5-a433-c9a50eeca1fc","Type":"ContainerStarted","Data":"9e7674a3a2e0262a2e4f86307acbce752fb0886cb8c874dfda7322f529680c99"} Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.941286 4750 generic.go:334] "Generic (PLEG): container finished" podID="06088b9b-c876-4cf1-978e-3aae41970e58" containerID="dcbd4793d502ff2bd54c8fdb312c0c7255764abeebb2cce77edb68d4b20b22ff" exitCode=0 Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.941385 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66h4d" event={"ID":"06088b9b-c876-4cf1-978e-3aae41970e58","Type":"ContainerDied","Data":"dcbd4793d502ff2bd54c8fdb312c0c7255764abeebb2cce77edb68d4b20b22ff"} Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.941416 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66h4d" event={"ID":"06088b9b-c876-4cf1-978e-3aae41970e58","Type":"ContainerStarted","Data":"fa8af6d2b00757cf9d56715ef86694b04cc725725cf5714842a7eefe95443f4b"} Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.943349 4750 generic.go:334] "Generic (PLEG): container finished" podID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerID="f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491" exitCode=0 Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.943378 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsrwr" event={"ID":"cdd878ef-1b79-49cd-a6bb-b756449fc7d8","Type":"ContainerDied","Data":"f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491"} Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.943398 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsrwr" event={"ID":"cdd878ef-1b79-49cd-a6bb-b756449fc7d8","Type":"ContainerStarted","Data":"002bb66edcf6f6d7dddf4733fd8e55e8611d2cf193cb2ebb587b044fe496d820"} Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.969385 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wtbl2" podStartSLOduration=2.4844657359999998 podStartE2EDuration="4.969370647s" podCreationTimestamp="2026-03-09 18:33:24 +0000 UTC" firstStartedPulling="2026-03-09 18:33:25.906467205 +0000 UTC m=+487.248939613" lastFinishedPulling="2026-03-09 18:33:28.391372116 +0000 UTC m=+489.733844524" observedRunningTime="2026-03-09 18:33:28.968003989 +0000 UTC m=+490.310476397" watchObservedRunningTime="2026-03-09 18:33:28.969370647 +0000 UTC m=+490.311843045" Mar 09 18:33:28 crc kubenswrapper[4750]: I0309 18:33:28.997386 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-55x9g" podStartSLOduration=1.403321863 podStartE2EDuration="3.997362201s" podCreationTimestamp="2026-03-09 18:33:25 +0000 UTC" firstStartedPulling="2026-03-09 18:33:25.900782826 +0000 UTC m=+487.243255234" lastFinishedPulling="2026-03-09 18:33:28.494823134 +0000 UTC m=+489.837295572" observedRunningTime="2026-03-09 18:33:28.987533436 +0000 UTC m=+490.330005834" watchObservedRunningTime="2026-03-09 18:33:28.997362201 +0000 UTC m=+490.339834599" Mar 09 18:33:29 crc kubenswrapper[4750]: I0309 18:33:29.951517 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsrwr" event={"ID":"cdd878ef-1b79-49cd-a6bb-b756449fc7d8","Type":"ContainerStarted","Data":"b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858"} Mar 09 18:33:29 crc kubenswrapper[4750]: I0309 18:33:29.955185 4750 generic.go:334] "Generic (PLEG): container finished" podID="06088b9b-c876-4cf1-978e-3aae41970e58" containerID="33f80dca00e0d0062108c1e127f90ccf5b3246edf2c504f3d6a3938b76a69e0f" exitCode=0 Mar 09 18:33:29 crc kubenswrapper[4750]: I0309 18:33:29.956377 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66h4d" event={"ID":"06088b9b-c876-4cf1-978e-3aae41970e58","Type":"ContainerDied","Data":"33f80dca00e0d0062108c1e127f90ccf5b3246edf2c504f3d6a3938b76a69e0f"} Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.322890 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" podUID="74eac05b-f613-45a9-b8fa-1eb83e8c04b4" containerName="registry" containerID="cri-o://d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc" gracePeriod=30 Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.743657 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.868307 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsfzt\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-kube-api-access-lsfzt\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.868423 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-tls\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.869772 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.869825 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-trusted-ca\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.869852 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-bound-sa-token\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.869894 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-ca-trust-extracted\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.869936 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-installation-pull-secrets\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.869975 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-certificates\") pod \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\" (UID: \"74eac05b-f613-45a9-b8fa-1eb83e8c04b4\") " Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.871289 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.871400 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.878508 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.878723 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.886079 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-kube-api-access-lsfzt" (OuterVolumeSpecName: "kube-api-access-lsfzt") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "kube-api-access-lsfzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.888328 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.891490 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.896363 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "74eac05b-f613-45a9-b8fa-1eb83e8c04b4" (UID: "74eac05b-f613-45a9-b8fa-1eb83e8c04b4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.964828 4750 generic.go:334] "Generic (PLEG): container finished" podID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerID="b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858" exitCode=0 Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.964947 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsrwr" event={"ID":"cdd878ef-1b79-49cd-a6bb-b756449fc7d8","Type":"ContainerDied","Data":"b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858"} Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.967448 4750 generic.go:334] "Generic (PLEG): container finished" podID="74eac05b-f613-45a9-b8fa-1eb83e8c04b4" containerID="d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc" exitCode=0 Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.967540 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" event={"ID":"74eac05b-f613-45a9-b8fa-1eb83e8c04b4","Type":"ContainerDied","Data":"d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc"} Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.967612 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" event={"ID":"74eac05b-f613-45a9-b8fa-1eb83e8c04b4","Type":"ContainerDied","Data":"ed312d0f39a71a1a8cfd66600a36b9b3443f0c153fcb647cdae15d549afe1416"} Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.967671 4750 scope.go:117] "RemoveContainer" containerID="d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.967506 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qmbbg" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.974474 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.974998 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.975152 4750 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.975254 4750 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.975347 4750 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.975432 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsfzt\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-kube-api-access-lsfzt\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.975516 4750 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/74eac05b-f613-45a9-b8fa-1eb83e8c04b4-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.977085 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-66h4d" event={"ID":"06088b9b-c876-4cf1-978e-3aae41970e58","Type":"ContainerStarted","Data":"4ac58cec1b00520f6bb1228d90113c9c8309e79c84c3149e732d1c319e369c4e"} Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.987251 4750 scope.go:117] "RemoveContainer" containerID="d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc" Mar 09 18:33:30 crc kubenswrapper[4750]: E0309 18:33:30.988176 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc\": container with ID starting with d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc not found: ID does not exist" containerID="d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc" Mar 09 18:33:30 crc kubenswrapper[4750]: I0309 18:33:30.988233 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc"} err="failed to get container status \"d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc\": rpc error: code = NotFound desc = could not find container \"d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc\": container with ID starting with d9f3fb4ba7630b5317499ed747665694b8b134384054fe306cd2e68c2b6563fc not found: ID does not exist" Mar 09 18:33:31 crc kubenswrapper[4750]: I0309 18:33:31.019864 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-66h4d" podStartSLOduration=2.567333198 podStartE2EDuration="4.019839678s" podCreationTimestamp="2026-03-09 18:33:27 +0000 UTC" firstStartedPulling="2026-03-09 18:33:28.945359135 +0000 UTC m=+490.287831563" lastFinishedPulling="2026-03-09 18:33:30.397865645 +0000 UTC m=+491.740338043" observedRunningTime="2026-03-09 18:33:31.018001217 +0000 UTC m=+492.360473615" watchObservedRunningTime="2026-03-09 18:33:31.019839678 +0000 UTC m=+492.362312076" Mar 09 18:33:31 crc kubenswrapper[4750]: I0309 18:33:31.035155 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qmbbg"] Mar 09 18:33:31 crc kubenswrapper[4750]: I0309 18:33:31.038918 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qmbbg"] Mar 09 18:33:31 crc kubenswrapper[4750]: I0309 18:33:31.387574 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74eac05b-f613-45a9-b8fa-1eb83e8c04b4" path="/var/lib/kubelet/pods/74eac05b-f613-45a9-b8fa-1eb83e8c04b4/volumes" Mar 09 18:33:31 crc kubenswrapper[4750]: I0309 18:33:31.987646 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsrwr" event={"ID":"cdd878ef-1b79-49cd-a6bb-b756449fc7d8","Type":"ContainerStarted","Data":"a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784"} Mar 09 18:33:32 crc kubenswrapper[4750]: I0309 18:33:32.010931 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qsrwr" podStartSLOduration=2.52988644 podStartE2EDuration="5.010907492s" podCreationTimestamp="2026-03-09 18:33:27 +0000 UTC" firstStartedPulling="2026-03-09 18:33:28.945330164 +0000 UTC m=+490.287802562" lastFinishedPulling="2026-03-09 18:33:31.426351216 +0000 UTC m=+492.768823614" observedRunningTime="2026-03-09 18:33:32.010718977 +0000 UTC m=+493.353191385" watchObservedRunningTime="2026-03-09 18:33:32.010907492 +0000 UTC m=+493.353379890" Mar 09 18:33:35 crc kubenswrapper[4750]: I0309 18:33:35.178776 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:35 crc kubenswrapper[4750]: I0309 18:33:35.179231 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:35 crc kubenswrapper[4750]: I0309 18:33:35.241937 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:35 crc kubenswrapper[4750]: I0309 18:33:35.367325 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:35 crc kubenswrapper[4750]: I0309 18:33:35.367820 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:35 crc kubenswrapper[4750]: I0309 18:33:35.434081 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:36 crc kubenswrapper[4750]: I0309 18:33:36.063670 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 18:33:36 crc kubenswrapper[4750]: I0309 18:33:36.076824 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-55x9g" Mar 09 18:33:37 crc kubenswrapper[4750]: I0309 18:33:37.652053 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:37 crc kubenswrapper[4750]: I0309 18:33:37.652437 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:37 crc kubenswrapper[4750]: I0309 18:33:37.696355 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:37 crc kubenswrapper[4750]: I0309 18:33:37.813116 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:37 crc kubenswrapper[4750]: I0309 18:33:37.813182 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:38 crc kubenswrapper[4750]: I0309 18:33:38.077824 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-66h4d" Mar 09 18:33:38 crc kubenswrapper[4750]: I0309 18:33:38.864356 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qsrwr" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="registry-server" probeResult="failure" output=< Mar 09 18:33:38 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 18:33:38 crc kubenswrapper[4750]: > Mar 09 18:33:47 crc kubenswrapper[4750]: I0309 18:33:47.855081 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:47 crc kubenswrapper[4750]: I0309 18:33:47.914558 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 18:33:51 crc kubenswrapper[4750]: I0309 18:33:51.744112 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:33:51 crc kubenswrapper[4750]: I0309 18:33:51.744716 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:33:51 crc kubenswrapper[4750]: I0309 18:33:51.744788 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:33:51 crc kubenswrapper[4750]: I0309 18:33:51.745674 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0818ed3c5ca0c09708e66182b106417995a2e2cc5b7ae8c6dec5d1157e85abf4"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:33:51 crc kubenswrapper[4750]: I0309 18:33:51.745752 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://0818ed3c5ca0c09708e66182b106417995a2e2cc5b7ae8c6dec5d1157e85abf4" gracePeriod=600 Mar 09 18:33:52 crc kubenswrapper[4750]: I0309 18:33:52.133163 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="0818ed3c5ca0c09708e66182b106417995a2e2cc5b7ae8c6dec5d1157e85abf4" exitCode=0 Mar 09 18:33:52 crc kubenswrapper[4750]: I0309 18:33:52.133271 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"0818ed3c5ca0c09708e66182b106417995a2e2cc5b7ae8c6dec5d1157e85abf4"} Mar 09 18:33:52 crc kubenswrapper[4750]: I0309 18:33:52.133696 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"dbc54b035b270a840e32e83b710affea17c5d6abaf9f561608508bb75a7047ba"} Mar 09 18:33:52 crc kubenswrapper[4750]: I0309 18:33:52.133730 4750 scope.go:117] "RemoveContainer" containerID="d87860fe9b21a075c92be2fd7298ea4754298cad73bb6a67ffa78db1c8171e26" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.151499 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551354-fpht2"] Mar 09 18:34:00 crc kubenswrapper[4750]: E0309 18:34:00.152344 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74eac05b-f613-45a9-b8fa-1eb83e8c04b4" containerName="registry" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.152361 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="74eac05b-f613-45a9-b8fa-1eb83e8c04b4" containerName="registry" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.152473 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="74eac05b-f613-45a9-b8fa-1eb83e8c04b4" containerName="registry" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.152976 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551354-fpht2" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.155960 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.156180 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.156826 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.159242 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551354-fpht2"] Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.228476 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8mc6\" (UniqueName: \"kubernetes.io/projected/6c072a7f-6cb4-4857-9425-ca75503d6470-kube-api-access-k8mc6\") pod \"auto-csr-approver-29551354-fpht2\" (UID: \"6c072a7f-6cb4-4857-9425-ca75503d6470\") " pod="openshift-infra/auto-csr-approver-29551354-fpht2" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.329825 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mc6\" (UniqueName: \"kubernetes.io/projected/6c072a7f-6cb4-4857-9425-ca75503d6470-kube-api-access-k8mc6\") pod \"auto-csr-approver-29551354-fpht2\" (UID: \"6c072a7f-6cb4-4857-9425-ca75503d6470\") " pod="openshift-infra/auto-csr-approver-29551354-fpht2" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.355330 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8mc6\" (UniqueName: \"kubernetes.io/projected/6c072a7f-6cb4-4857-9425-ca75503d6470-kube-api-access-k8mc6\") pod \"auto-csr-approver-29551354-fpht2\" (UID: \"6c072a7f-6cb4-4857-9425-ca75503d6470\") " pod="openshift-infra/auto-csr-approver-29551354-fpht2" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.473427 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551354-fpht2" Mar 09 18:34:00 crc kubenswrapper[4750]: I0309 18:34:00.718702 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551354-fpht2"] Mar 09 18:34:01 crc kubenswrapper[4750]: I0309 18:34:01.201912 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551354-fpht2" event={"ID":"6c072a7f-6cb4-4857-9425-ca75503d6470","Type":"ContainerStarted","Data":"d1ed3bb1405a60cfa1b61a567d8ec12547d77bd1dce3ed8cd481e8e3a484a27e"} Mar 09 18:34:03 crc kubenswrapper[4750]: I0309 18:34:03.218047 4750 generic.go:334] "Generic (PLEG): container finished" podID="6c072a7f-6cb4-4857-9425-ca75503d6470" containerID="0956a3e4f74a7e5ce542709aef043452900c4a885aa3b666b24560e27d248940" exitCode=0 Mar 09 18:34:03 crc kubenswrapper[4750]: I0309 18:34:03.218172 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551354-fpht2" event={"ID":"6c072a7f-6cb4-4857-9425-ca75503d6470","Type":"ContainerDied","Data":"0956a3e4f74a7e5ce542709aef043452900c4a885aa3b666b24560e27d248940"} Mar 09 18:34:04 crc kubenswrapper[4750]: I0309 18:34:04.440341 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551354-fpht2" Mar 09 18:34:04 crc kubenswrapper[4750]: I0309 18:34:04.588614 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8mc6\" (UniqueName: \"kubernetes.io/projected/6c072a7f-6cb4-4857-9425-ca75503d6470-kube-api-access-k8mc6\") pod \"6c072a7f-6cb4-4857-9425-ca75503d6470\" (UID: \"6c072a7f-6cb4-4857-9425-ca75503d6470\") " Mar 09 18:34:04 crc kubenswrapper[4750]: I0309 18:34:04.595174 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c072a7f-6cb4-4857-9425-ca75503d6470-kube-api-access-k8mc6" (OuterVolumeSpecName: "kube-api-access-k8mc6") pod "6c072a7f-6cb4-4857-9425-ca75503d6470" (UID: "6c072a7f-6cb4-4857-9425-ca75503d6470"). InnerVolumeSpecName "kube-api-access-k8mc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:34:04 crc kubenswrapper[4750]: I0309 18:34:04.689961 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8mc6\" (UniqueName: \"kubernetes.io/projected/6c072a7f-6cb4-4857-9425-ca75503d6470-kube-api-access-k8mc6\") on node \"crc\" DevicePath \"\"" Mar 09 18:34:05 crc kubenswrapper[4750]: I0309 18:34:05.234480 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551354-fpht2" event={"ID":"6c072a7f-6cb4-4857-9425-ca75503d6470","Type":"ContainerDied","Data":"d1ed3bb1405a60cfa1b61a567d8ec12547d77bd1dce3ed8cd481e8e3a484a27e"} Mar 09 18:34:05 crc kubenswrapper[4750]: I0309 18:34:05.234538 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1ed3bb1405a60cfa1b61a567d8ec12547d77bd1dce3ed8cd481e8e3a484a27e" Mar 09 18:34:05 crc kubenswrapper[4750]: I0309 18:34:05.234696 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551354-fpht2" Mar 09 18:34:05 crc kubenswrapper[4750]: I0309 18:34:05.501388 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551348-g9vjb"] Mar 09 18:34:05 crc kubenswrapper[4750]: I0309 18:34:05.505881 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551348-g9vjb"] Mar 09 18:34:07 crc kubenswrapper[4750]: I0309 18:34:07.386282 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a08fcb79-42d8-47c2-940e-92c4ff4f2f61" path="/var/lib/kubelet/pods/a08fcb79-42d8-47c2-940e-92c4ff4f2f61/volumes" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.157898 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551356-w5mg4"] Mar 09 18:36:00 crc kubenswrapper[4750]: E0309 18:36:00.159548 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c072a7f-6cb4-4857-9425-ca75503d6470" containerName="oc" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.159581 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c072a7f-6cb4-4857-9425-ca75503d6470" containerName="oc" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.159816 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c072a7f-6cb4-4857-9425-ca75503d6470" containerName="oc" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.160503 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551356-w5mg4" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.163401 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.164825 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.165469 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.175469 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551356-w5mg4"] Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.214170 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkwmw\" (UniqueName: \"kubernetes.io/projected/957dd1c4-3eac-4206-b18e-a337e54ca850-kube-api-access-nkwmw\") pod \"auto-csr-approver-29551356-w5mg4\" (UID: \"957dd1c4-3eac-4206-b18e-a337e54ca850\") " pod="openshift-infra/auto-csr-approver-29551356-w5mg4" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.316107 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkwmw\" (UniqueName: \"kubernetes.io/projected/957dd1c4-3eac-4206-b18e-a337e54ca850-kube-api-access-nkwmw\") pod \"auto-csr-approver-29551356-w5mg4\" (UID: \"957dd1c4-3eac-4206-b18e-a337e54ca850\") " pod="openshift-infra/auto-csr-approver-29551356-w5mg4" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.355685 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkwmw\" (UniqueName: \"kubernetes.io/projected/957dd1c4-3eac-4206-b18e-a337e54ca850-kube-api-access-nkwmw\") pod \"auto-csr-approver-29551356-w5mg4\" (UID: \"957dd1c4-3eac-4206-b18e-a337e54ca850\") " pod="openshift-infra/auto-csr-approver-29551356-w5mg4" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.498111 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551356-w5mg4" Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.780677 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551356-w5mg4"] Mar 09 18:36:00 crc kubenswrapper[4750]: I0309 18:36:00.794360 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 18:36:01 crc kubenswrapper[4750]: I0309 18:36:01.170015 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551356-w5mg4" event={"ID":"957dd1c4-3eac-4206-b18e-a337e54ca850","Type":"ContainerStarted","Data":"163c56b213cd313c806af7252672a45906290cbcb5e43c0dda697582d7a74f8d"} Mar 09 18:36:03 crc kubenswrapper[4750]: I0309 18:36:03.192103 4750 generic.go:334] "Generic (PLEG): container finished" podID="957dd1c4-3eac-4206-b18e-a337e54ca850" containerID="e6f7f0768d424f3fce16f78de9d11dfd4696c27337a61d96238e62647e39a3ae" exitCode=0 Mar 09 18:36:03 crc kubenswrapper[4750]: I0309 18:36:03.192211 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551356-w5mg4" event={"ID":"957dd1c4-3eac-4206-b18e-a337e54ca850","Type":"ContainerDied","Data":"e6f7f0768d424f3fce16f78de9d11dfd4696c27337a61d96238e62647e39a3ae"} Mar 09 18:36:04 crc kubenswrapper[4750]: I0309 18:36:04.507467 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551356-w5mg4" Mar 09 18:36:04 crc kubenswrapper[4750]: I0309 18:36:04.579999 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkwmw\" (UniqueName: \"kubernetes.io/projected/957dd1c4-3eac-4206-b18e-a337e54ca850-kube-api-access-nkwmw\") pod \"957dd1c4-3eac-4206-b18e-a337e54ca850\" (UID: \"957dd1c4-3eac-4206-b18e-a337e54ca850\") " Mar 09 18:36:04 crc kubenswrapper[4750]: I0309 18:36:04.589186 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/957dd1c4-3eac-4206-b18e-a337e54ca850-kube-api-access-nkwmw" (OuterVolumeSpecName: "kube-api-access-nkwmw") pod "957dd1c4-3eac-4206-b18e-a337e54ca850" (UID: "957dd1c4-3eac-4206-b18e-a337e54ca850"). InnerVolumeSpecName "kube-api-access-nkwmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:36:04 crc kubenswrapper[4750]: I0309 18:36:04.681218 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkwmw\" (UniqueName: \"kubernetes.io/projected/957dd1c4-3eac-4206-b18e-a337e54ca850-kube-api-access-nkwmw\") on node \"crc\" DevicePath \"\"" Mar 09 18:36:05 crc kubenswrapper[4750]: I0309 18:36:05.216398 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551356-w5mg4" event={"ID":"957dd1c4-3eac-4206-b18e-a337e54ca850","Type":"ContainerDied","Data":"163c56b213cd313c806af7252672a45906290cbcb5e43c0dda697582d7a74f8d"} Mar 09 18:36:05 crc kubenswrapper[4750]: I0309 18:36:05.216466 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="163c56b213cd313c806af7252672a45906290cbcb5e43c0dda697582d7a74f8d" Mar 09 18:36:05 crc kubenswrapper[4750]: I0309 18:36:05.216559 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551356-w5mg4" Mar 09 18:36:05 crc kubenswrapper[4750]: I0309 18:36:05.582880 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551350-vmx4v"] Mar 09 18:36:05 crc kubenswrapper[4750]: I0309 18:36:05.590539 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551350-vmx4v"] Mar 09 18:36:07 crc kubenswrapper[4750]: I0309 18:36:07.386877 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f670b4-f234-4c50-beb3-5ae4d0492d3f" path="/var/lib/kubelet/pods/a5f670b4-f234-4c50-beb3-5ae4d0492d3f/volumes" Mar 09 18:36:19 crc kubenswrapper[4750]: I0309 18:36:19.732821 4750 scope.go:117] "RemoveContainer" containerID="0ca0155e7ea9f0108cf9a825a6a8d9b2a6513ff48fa06e8ff5cde596fe50e039" Mar 09 18:36:19 crc kubenswrapper[4750]: I0309 18:36:19.763809 4750 scope.go:117] "RemoveContainer" containerID="06cecdc078351cbf290c74ea9df45ed62c823161e4a770caa783c20d20debcb0" Mar 09 18:36:19 crc kubenswrapper[4750]: I0309 18:36:19.821391 4750 scope.go:117] "RemoveContainer" containerID="4251eb29c736506ce6e889790879763b504a1939bdae6db77b810fae289f67a4" Mar 09 18:36:21 crc kubenswrapper[4750]: I0309 18:36:21.744126 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:36:21 crc kubenswrapper[4750]: I0309 18:36:21.744336 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:36:51 crc kubenswrapper[4750]: I0309 18:36:51.743487 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:36:51 crc kubenswrapper[4750]: I0309 18:36:51.744535 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:37:21 crc kubenswrapper[4750]: I0309 18:37:21.743845 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:37:21 crc kubenswrapper[4750]: I0309 18:37:21.745832 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:37:21 crc kubenswrapper[4750]: I0309 18:37:21.745939 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:37:21 crc kubenswrapper[4750]: I0309 18:37:21.746971 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbc54b035b270a840e32e83b710affea17c5d6abaf9f561608508bb75a7047ba"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:37:21 crc kubenswrapper[4750]: I0309 18:37:21.747080 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://dbc54b035b270a840e32e83b710affea17c5d6abaf9f561608508bb75a7047ba" gracePeriod=600 Mar 09 18:37:22 crc kubenswrapper[4750]: I0309 18:37:22.875149 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="dbc54b035b270a840e32e83b710affea17c5d6abaf9f561608508bb75a7047ba" exitCode=0 Mar 09 18:37:22 crc kubenswrapper[4750]: I0309 18:37:22.875232 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"dbc54b035b270a840e32e83b710affea17c5d6abaf9f561608508bb75a7047ba"} Mar 09 18:37:22 crc kubenswrapper[4750]: I0309 18:37:22.876030 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"d522a3b191af7d8f923266bd379a6ce4a390d5d0273a9db4e49eaf7ceef3304f"} Mar 09 18:37:22 crc kubenswrapper[4750]: I0309 18:37:22.876068 4750 scope.go:117] "RemoveContainer" containerID="0818ed3c5ca0c09708e66182b106417995a2e2cc5b7ae8c6dec5d1157e85abf4" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.154179 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551358-gqjcz"] Mar 09 18:38:00 crc kubenswrapper[4750]: E0309 18:38:00.155425 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957dd1c4-3eac-4206-b18e-a337e54ca850" containerName="oc" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.155458 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="957dd1c4-3eac-4206-b18e-a337e54ca850" containerName="oc" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.155728 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="957dd1c4-3eac-4206-b18e-a337e54ca850" containerName="oc" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.156680 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.161056 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.161107 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551358-gqjcz"] Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.161318 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.161529 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.346917 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qc57\" (UniqueName: \"kubernetes.io/projected/8f9c03a3-e6ea-411a-a9b7-99a98c41492f-kube-api-access-4qc57\") pod \"auto-csr-approver-29551358-gqjcz\" (UID: \"8f9c03a3-e6ea-411a-a9b7-99a98c41492f\") " pod="openshift-infra/auto-csr-approver-29551358-gqjcz" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.448437 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qc57\" (UniqueName: \"kubernetes.io/projected/8f9c03a3-e6ea-411a-a9b7-99a98c41492f-kube-api-access-4qc57\") pod \"auto-csr-approver-29551358-gqjcz\" (UID: \"8f9c03a3-e6ea-411a-a9b7-99a98c41492f\") " pod="openshift-infra/auto-csr-approver-29551358-gqjcz" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.477218 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qc57\" (UniqueName: \"kubernetes.io/projected/8f9c03a3-e6ea-411a-a9b7-99a98c41492f-kube-api-access-4qc57\") pod \"auto-csr-approver-29551358-gqjcz\" (UID: \"8f9c03a3-e6ea-411a-a9b7-99a98c41492f\") " pod="openshift-infra/auto-csr-approver-29551358-gqjcz" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.493774 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" Mar 09 18:38:00 crc kubenswrapper[4750]: I0309 18:38:00.756554 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551358-gqjcz"] Mar 09 18:38:00 crc kubenswrapper[4750]: W0309 18:38:00.768844 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f9c03a3_e6ea_411a_a9b7_99a98c41492f.slice/crio-e3376c47da48837ccb8819fcca180a66bdefb37857913d0200f889f3ed81bda1 WatchSource:0}: Error finding container e3376c47da48837ccb8819fcca180a66bdefb37857913d0200f889f3ed81bda1: Status 404 returned error can't find the container with id e3376c47da48837ccb8819fcca180a66bdefb37857913d0200f889f3ed81bda1 Mar 09 18:38:01 crc kubenswrapper[4750]: I0309 18:38:01.173256 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" event={"ID":"8f9c03a3-e6ea-411a-a9b7-99a98c41492f","Type":"ContainerStarted","Data":"e3376c47da48837ccb8819fcca180a66bdefb37857913d0200f889f3ed81bda1"} Mar 09 18:38:02 crc kubenswrapper[4750]: I0309 18:38:02.184084 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" event={"ID":"8f9c03a3-e6ea-411a-a9b7-99a98c41492f","Type":"ContainerStarted","Data":"e7928fd87977ab59a664baed25da0b8c02b15737bceb8fc18cace73eddd3f19a"} Mar 09 18:38:02 crc kubenswrapper[4750]: I0309 18:38:02.202697 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" podStartSLOduration=1.118803087 podStartE2EDuration="2.202670061s" podCreationTimestamp="2026-03-09 18:38:00 +0000 UTC" firstStartedPulling="2026-03-09 18:38:00.771580589 +0000 UTC m=+762.114052987" lastFinishedPulling="2026-03-09 18:38:01.855447563 +0000 UTC m=+763.197919961" observedRunningTime="2026-03-09 18:38:02.201080827 +0000 UTC m=+763.543553235" watchObservedRunningTime="2026-03-09 18:38:02.202670061 +0000 UTC m=+763.545142479" Mar 09 18:38:03 crc kubenswrapper[4750]: I0309 18:38:03.215558 4750 generic.go:334] "Generic (PLEG): container finished" podID="8f9c03a3-e6ea-411a-a9b7-99a98c41492f" containerID="e7928fd87977ab59a664baed25da0b8c02b15737bceb8fc18cace73eddd3f19a" exitCode=0 Mar 09 18:38:03 crc kubenswrapper[4750]: I0309 18:38:03.215699 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" event={"ID":"8f9c03a3-e6ea-411a-a9b7-99a98c41492f","Type":"ContainerDied","Data":"e7928fd87977ab59a664baed25da0b8c02b15737bceb8fc18cace73eddd3f19a"} Mar 09 18:38:04 crc kubenswrapper[4750]: I0309 18:38:04.528421 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" Mar 09 18:38:04 crc kubenswrapper[4750]: I0309 18:38:04.709987 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qc57\" (UniqueName: \"kubernetes.io/projected/8f9c03a3-e6ea-411a-a9b7-99a98c41492f-kube-api-access-4qc57\") pod \"8f9c03a3-e6ea-411a-a9b7-99a98c41492f\" (UID: \"8f9c03a3-e6ea-411a-a9b7-99a98c41492f\") " Mar 09 18:38:04 crc kubenswrapper[4750]: I0309 18:38:04.718051 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9c03a3-e6ea-411a-a9b7-99a98c41492f-kube-api-access-4qc57" (OuterVolumeSpecName: "kube-api-access-4qc57") pod "8f9c03a3-e6ea-411a-a9b7-99a98c41492f" (UID: "8f9c03a3-e6ea-411a-a9b7-99a98c41492f"). InnerVolumeSpecName "kube-api-access-4qc57". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:38:04 crc kubenswrapper[4750]: I0309 18:38:04.811604 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qc57\" (UniqueName: \"kubernetes.io/projected/8f9c03a3-e6ea-411a-a9b7-99a98c41492f-kube-api-access-4qc57\") on node \"crc\" DevicePath \"\"" Mar 09 18:38:05 crc kubenswrapper[4750]: I0309 18:38:05.253645 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" event={"ID":"8f9c03a3-e6ea-411a-a9b7-99a98c41492f","Type":"ContainerDied","Data":"e3376c47da48837ccb8819fcca180a66bdefb37857913d0200f889f3ed81bda1"} Mar 09 18:38:05 crc kubenswrapper[4750]: I0309 18:38:05.253712 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3376c47da48837ccb8819fcca180a66bdefb37857913d0200f889f3ed81bda1" Mar 09 18:38:05 crc kubenswrapper[4750]: I0309 18:38:05.253801 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551358-gqjcz" Mar 09 18:38:05 crc kubenswrapper[4750]: I0309 18:38:05.291911 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551352-mqsjs"] Mar 09 18:38:05 crc kubenswrapper[4750]: I0309 18:38:05.295804 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551352-mqsjs"] Mar 09 18:38:05 crc kubenswrapper[4750]: I0309 18:38:05.385707 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="160b58c3-8419-4f38-8aae-e02628170cb7" path="/var/lib/kubelet/pods/160b58c3-8419-4f38-8aae-e02628170cb7/volumes" Mar 09 18:38:19 crc kubenswrapper[4750]: I0309 18:38:19.946761 4750 scope.go:117] "RemoveContainer" containerID="7cd34d7145390006e08fbc6ead05527af5e8be76b668219103e4202e44c5a96c" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.017167 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6"] Mar 09 18:39:09 crc kubenswrapper[4750]: E0309 18:39:09.018153 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9c03a3-e6ea-411a-a9b7-99a98c41492f" containerName="oc" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.018169 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9c03a3-e6ea-411a-a9b7-99a98c41492f" containerName="oc" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.018284 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9c03a3-e6ea-411a-a9b7-99a98c41492f" containerName="oc" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.018716 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.020603 4750 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-scdpg" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.021063 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.033940 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6"] Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.039258 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.043718 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-xxfrp"] Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.044663 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-xxfrp" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.047142 4750 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-mbtm9" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.066918 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-xxfrp"] Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.078791 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8pf5j"] Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.079958 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.083542 4750 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4z4mc" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.091374 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8pf5j"] Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.126821 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbphc\" (UniqueName: \"kubernetes.io/projected/45e5d4bf-93bf-4bea-9ee4-656c3c4f8077-kube-api-access-cbphc\") pod \"cert-manager-cainjector-cf98fcc89-2k6r6\" (UID: \"45e5d4bf-93bf-4bea-9ee4-656c3c4f8077\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.228087 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s42dq\" (UniqueName: \"kubernetes.io/projected/d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c-kube-api-access-s42dq\") pod \"cert-manager-858654f9db-xxfrp\" (UID: \"d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c\") " pod="cert-manager/cert-manager-858654f9db-xxfrp" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.228169 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbphc\" (UniqueName: \"kubernetes.io/projected/45e5d4bf-93bf-4bea-9ee4-656c3c4f8077-kube-api-access-cbphc\") pod \"cert-manager-cainjector-cf98fcc89-2k6r6\" (UID: \"45e5d4bf-93bf-4bea-9ee4-656c3c4f8077\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.228215 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqb5z\" (UniqueName: \"kubernetes.io/projected/b7eaca9e-324d-4934-bcac-d8a6034c94f6-kube-api-access-fqb5z\") pod \"cert-manager-webhook-687f57d79b-8pf5j\" (UID: \"b7eaca9e-324d-4934-bcac-d8a6034c94f6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.257246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbphc\" (UniqueName: \"kubernetes.io/projected/45e5d4bf-93bf-4bea-9ee4-656c3c4f8077-kube-api-access-cbphc\") pod \"cert-manager-cainjector-cf98fcc89-2k6r6\" (UID: \"45e5d4bf-93bf-4bea-9ee4-656c3c4f8077\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.329892 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s42dq\" (UniqueName: \"kubernetes.io/projected/d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c-kube-api-access-s42dq\") pod \"cert-manager-858654f9db-xxfrp\" (UID: \"d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c\") " pod="cert-manager/cert-manager-858654f9db-xxfrp" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.329982 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqb5z\" (UniqueName: \"kubernetes.io/projected/b7eaca9e-324d-4934-bcac-d8a6034c94f6-kube-api-access-fqb5z\") pod \"cert-manager-webhook-687f57d79b-8pf5j\" (UID: \"b7eaca9e-324d-4934-bcac-d8a6034c94f6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.342710 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.348196 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s42dq\" (UniqueName: \"kubernetes.io/projected/d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c-kube-api-access-s42dq\") pod \"cert-manager-858654f9db-xxfrp\" (UID: \"d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c\") " pod="cert-manager/cert-manager-858654f9db-xxfrp" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.358019 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqb5z\" (UniqueName: \"kubernetes.io/projected/b7eaca9e-324d-4934-bcac-d8a6034c94f6-kube-api-access-fqb5z\") pod \"cert-manager-webhook-687f57d79b-8pf5j\" (UID: \"b7eaca9e-324d-4934-bcac-d8a6034c94f6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.362341 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-xxfrp" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.396423 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.827042 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6"] Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.906767 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-xxfrp"] Mar 09 18:39:09 crc kubenswrapper[4750]: I0309 18:39:09.907748 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8pf5j"] Mar 09 18:39:09 crc kubenswrapper[4750]: W0309 18:39:09.910293 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4fb6a3f_ccb3_4df8_b8f7_2d3b3295448c.slice/crio-d9d2a2bb702b54474ec2119f0a2ad9682cc80315db609ad0a515f040838ee569 WatchSource:0}: Error finding container d9d2a2bb702b54474ec2119f0a2ad9682cc80315db609ad0a515f040838ee569: Status 404 returned error can't find the container with id d9d2a2bb702b54474ec2119f0a2ad9682cc80315db609ad0a515f040838ee569 Mar 09 18:39:09 crc kubenswrapper[4750]: W0309 18:39:09.911714 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7eaca9e_324d_4934_bcac_d8a6034c94f6.slice/crio-4be7302e2a389adf83b882ee5641a2bbdfcb8e5e5ccfe5ed8b0731cf35008252 WatchSource:0}: Error finding container 4be7302e2a389adf83b882ee5641a2bbdfcb8e5e5ccfe5ed8b0731cf35008252: Status 404 returned error can't find the container with id 4be7302e2a389adf83b882ee5641a2bbdfcb8e5e5ccfe5ed8b0731cf35008252 Mar 09 18:39:10 crc kubenswrapper[4750]: I0309 18:39:10.743863 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" event={"ID":"45e5d4bf-93bf-4bea-9ee4-656c3c4f8077","Type":"ContainerStarted","Data":"4c99dbe5206a8ee118a6e12c0de29df0e9a2d9b574fb291a5de8dcbfbd6a9c99"} Mar 09 18:39:10 crc kubenswrapper[4750]: I0309 18:39:10.746758 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-xxfrp" event={"ID":"d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c","Type":"ContainerStarted","Data":"d9d2a2bb702b54474ec2119f0a2ad9682cc80315db609ad0a515f040838ee569"} Mar 09 18:39:10 crc kubenswrapper[4750]: I0309 18:39:10.748842 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" event={"ID":"b7eaca9e-324d-4934-bcac-d8a6034c94f6","Type":"ContainerStarted","Data":"4be7302e2a389adf83b882ee5641a2bbdfcb8e5e5ccfe5ed8b0731cf35008252"} Mar 09 18:39:14 crc kubenswrapper[4750]: I0309 18:39:14.787117 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" event={"ID":"b7eaca9e-324d-4934-bcac-d8a6034c94f6","Type":"ContainerStarted","Data":"c4a579dcd37eccc89b59dbd3482bc3ec4a039450f18bac817adb17aedf8ad9a2"} Mar 09 18:39:14 crc kubenswrapper[4750]: I0309 18:39:14.788910 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" Mar 09 18:39:14 crc kubenswrapper[4750]: I0309 18:39:14.790785 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" event={"ID":"45e5d4bf-93bf-4bea-9ee4-656c3c4f8077","Type":"ContainerStarted","Data":"3a3fa9e4985381cee598dc58b2fd2723e49b10f684f555b7ef9de80dfc725fc5"} Mar 09 18:39:14 crc kubenswrapper[4750]: I0309 18:39:14.797772 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-xxfrp" event={"ID":"d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c","Type":"ContainerStarted","Data":"d4f6033a927c0028ecbac9d53bea1ba4089e3cf70aa7db9983d7e711a26ad852"} Mar 09 18:39:14 crc kubenswrapper[4750]: I0309 18:39:14.814145 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" podStartSLOduration=2.077661656 podStartE2EDuration="5.814121523s" podCreationTimestamp="2026-03-09 18:39:09 +0000 UTC" firstStartedPulling="2026-03-09 18:39:09.914083705 +0000 UTC m=+831.256556113" lastFinishedPulling="2026-03-09 18:39:13.650543582 +0000 UTC m=+834.993015980" observedRunningTime="2026-03-09 18:39:14.811152733 +0000 UTC m=+836.153625131" watchObservedRunningTime="2026-03-09 18:39:14.814121523 +0000 UTC m=+836.156593921" Mar 09 18:39:14 crc kubenswrapper[4750]: I0309 18:39:14.835735 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2k6r6" podStartSLOduration=2.096696719 podStartE2EDuration="5.835707523s" podCreationTimestamp="2026-03-09 18:39:09 +0000 UTC" firstStartedPulling="2026-03-09 18:39:09.83806124 +0000 UTC m=+831.180533638" lastFinishedPulling="2026-03-09 18:39:13.577072004 +0000 UTC m=+834.919544442" observedRunningTime="2026-03-09 18:39:14.827882463 +0000 UTC m=+836.170354881" watchObservedRunningTime="2026-03-09 18:39:14.835707523 +0000 UTC m=+836.178179941" Mar 09 18:39:14 crc kubenswrapper[4750]: I0309 18:39:14.863395 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-xxfrp" podStartSLOduration=2.199882467 podStartE2EDuration="5.863370128s" podCreationTimestamp="2026-03-09 18:39:09 +0000 UTC" firstStartedPulling="2026-03-09 18:39:09.913587523 +0000 UTC m=+831.256059941" lastFinishedPulling="2026-03-09 18:39:13.577075174 +0000 UTC m=+834.919547602" observedRunningTime="2026-03-09 18:39:14.854283973 +0000 UTC m=+836.196756381" watchObservedRunningTime="2026-03-09 18:39:14.863370128 +0000 UTC m=+836.205842516" Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.802421 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lfwr6"] Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.803847 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-controller" containerID="cri-o://1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" gracePeriod=30 Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.803921 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="nbdb" containerID="cri-o://96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" gracePeriod=30 Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.804088 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-node" containerID="cri-o://a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" gracePeriod=30 Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.804176 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-acl-logging" containerID="cri-o://4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" gracePeriod=30 Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.804101 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="sbdb" containerID="cri-o://f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" gracePeriod=30 Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.804305 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="northd" containerID="cri-o://f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" gracePeriod=30 Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.804608 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" gracePeriod=30 Mar 09 18:39:18 crc kubenswrapper[4750]: I0309 18:39:18.847526 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" containerID="cri-o://1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" gracePeriod=30 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.164861 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/3.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.168465 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovn-acl-logging/0.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.169140 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovn-controller/0.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.169727 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183581 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-var-lib-openvswitch\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183690 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-ovn\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183724 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183766 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-config\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183747 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183798 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-netns\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183827 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-systemd-units\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183852 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-openvswitch\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183875 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-ovn-kubernetes\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183900 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-env-overrides\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.183907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.184011 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.184042 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.184031 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.184131 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/80b0e746-0802-4112-b77b-4db85dc85a9c-ovn-node-metrics-cert\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.184320 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.184601 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.184798 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185038 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-etc-openvswitch\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185093 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-script-lib\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185125 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-netd\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185148 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185186 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v5n4\" (UniqueName: \"kubernetes.io/projected/80b0e746-0802-4112-b77b-4db85dc85a9c-kube-api-access-7v5n4\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185211 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-slash\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185240 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-log-socket\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185273 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-node-log\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185238 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185261 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185330 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185337 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-log-socket" (OuterVolumeSpecName: "log-socket") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185302 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-bin\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185381 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-node-log" (OuterVolumeSpecName: "node-log") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185401 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-slash" (OuterVolumeSpecName: "host-slash") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185447 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-kubelet\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185502 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.185900 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186015 4750 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186049 4750 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-slash\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186063 4750 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-log-socket\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186075 4750 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-node-log\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186087 4750 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186101 4750 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186114 4750 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186126 4750 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186138 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186151 4750 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186164 4750 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186176 4750 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186190 4750 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186203 4750 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186216 4750 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186227 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/80b0e746-0802-4112-b77b-4db85dc85a9c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.186239 4750 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239484 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hx78v"] Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239796 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="northd" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239819 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="northd" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239844 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kubecfg-setup" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239854 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kubecfg-setup" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239867 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-ovn-metrics" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239878 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-ovn-metrics" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239892 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239902 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239916 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-acl-logging" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239924 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-acl-logging" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239937 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-node" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239946 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-node" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239955 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239964 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239973 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.239982 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.239992 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="sbdb" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240000 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="sbdb" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.240014 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240024 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.240037 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="nbdb" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240046 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="nbdb" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.240062 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240070 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.240081 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240089 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240229 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240240 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240251 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240260 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240270 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-node" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240306 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="nbdb" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240320 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240333 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="sbdb" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240342 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="northd" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240356 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovn-acl-logging" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240367 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="kube-rbac-proxy-ovn-metrics" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.240709 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerName="ovnkube-controller" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.409835 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.410423 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b0e746-0802-4112-b77b-4db85dc85a9c-kube-api-access-7v5n4" (OuterVolumeSpecName: "kube-api-access-7v5n4") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "kube-api-access-7v5n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.410947 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b0e746-0802-4112-b77b-4db85dc85a9c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.416918 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd\") pod \"80b0e746-0802-4112-b77b-4db85dc85a9c\" (UID: \"80b0e746-0802-4112-b77b-4db85dc85a9c\") " Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.419908 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "80b0e746-0802-4112-b77b-4db85dc85a9c" (UID: "80b0e746-0802-4112-b77b-4db85dc85a9c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.427688 4750 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/80b0e746-0802-4112-b77b-4db85dc85a9c-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.427750 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/80b0e746-0802-4112-b77b-4db85dc85a9c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.427773 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v5n4\" (UniqueName: \"kubernetes.io/projected/80b0e746-0802-4112-b77b-4db85dc85a9c-kube-api-access-7v5n4\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.437458 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-8pf5j" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.437977 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: E0309 18:39:19.539583 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80b0e746_0802_4112_b77b_4db85dc85a9c.slice/crio-eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93\": RecentStats: unable to find data in memory cache]" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.630487 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-var-lib-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.630969 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovn-node-metrics-cert\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.631070 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-ovn\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.631120 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfvmk\" (UniqueName: \"kubernetes.io/projected/ee99c41c-bd53-4dd7-be2b-a8916902a942-kube-api-access-lfvmk\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.631212 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.631257 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-systemd-units\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.631281 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-run-ovn-kubernetes\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.631674 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovnkube-script-lib\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632248 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-etc-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632284 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-cni-netd\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632411 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-slash\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632449 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632502 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-systemd\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632588 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-kubelet\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632624 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-log-socket\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632764 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-run-netns\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632808 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-env-overrides\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632908 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-node-log\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632937 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-cni-bin\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.632967 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovnkube-config\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734248 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-etc-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734336 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-cni-netd\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734381 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-slash\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734414 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734440 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-systemd\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734472 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-kubelet\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734497 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-log-socket\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734484 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-etc-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734521 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-run-netns\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734884 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-env-overrides\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734538 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-cni-netd\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734958 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-node-log\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735026 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-node-log\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734644 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735051 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-cni-bin\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735097 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-cni-bin\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734585 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-run-netns\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734582 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-kubelet\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735131 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovnkube-config\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734672 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-log-socket\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734652 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-systemd\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.734569 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-slash\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735357 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-var-lib-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735405 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovn-node-metrics-cert\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735458 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-var-lib-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735482 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-ovn\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735755 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfvmk\" (UniqueName: \"kubernetes.io/projected/ee99c41c-bd53-4dd7-be2b-a8916902a942-kube-api-access-lfvmk\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735794 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735813 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-ovn\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735858 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-systemd-units\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735827 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-systemd-units\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735889 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-env-overrides\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735918 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-run-ovn-kubernetes\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735954 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-host-run-ovn-kubernetes\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.735886 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee99c41c-bd53-4dd7-be2b-a8916902a942-run-openvswitch\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.736034 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovnkube-script-lib\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.737028 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovnkube-script-lib\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.737142 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovnkube-config\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.741703 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee99c41c-bd53-4dd7-be2b-a8916902a942-ovn-node-metrics-cert\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.755650 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfvmk\" (UniqueName: \"kubernetes.io/projected/ee99c41c-bd53-4dd7-be2b-a8916902a942-kube-api-access-lfvmk\") pod \"ovnkube-node-hx78v\" (UID: \"ee99c41c-bd53-4dd7-be2b-a8916902a942\") " pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.768554 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:19 crc kubenswrapper[4750]: W0309 18:39:19.800853 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee99c41c_bd53_4dd7_be2b_a8916902a942.slice/crio-9bcc60db49071fcce8d352c20b840bd05b05e38431a525abfef3989ca34efbd7 WatchSource:0}: Error finding container 9bcc60db49071fcce8d352c20b840bd05b05e38431a525abfef3989ca34efbd7: Status 404 returned error can't find the container with id 9bcc60db49071fcce8d352c20b840bd05b05e38431a525abfef3989ca34efbd7 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.844745 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/2.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.845469 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/1.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.845578 4750 generic.go:334] "Generic (PLEG): container finished" podID="fa652827-92a6-4544-b1b6-cba227f8e9f4" containerID="22b0048bade8f6e58ad502d87ccb4eefd6bb1e25e79954adaafc0157a254f540" exitCode=2 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.845663 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerDied","Data":"22b0048bade8f6e58ad502d87ccb4eefd6bb1e25e79954adaafc0157a254f540"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.845745 4750 scope.go:117] "RemoveContainer" containerID="2b557a5c154254066c248fe9b3faf35c5c413dd6dccd6b3cb6c9ce588f8fa64c" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.846145 4750 scope.go:117] "RemoveContainer" containerID="22b0048bade8f6e58ad502d87ccb4eefd6bb1e25e79954adaafc0157a254f540" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.849351 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovnkube-controller/3.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.852754 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovn-acl-logging/0.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.854903 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lfwr6_80b0e746-0802-4112-b77b-4db85dc85a9c/ovn-controller/0.log" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.858846 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" exitCode=0 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859000 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" exitCode=0 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859077 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" exitCode=0 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859168 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" exitCode=0 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859231 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" exitCode=0 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859284 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" exitCode=0 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859346 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" exitCode=143 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859407 4750 generic.go:334] "Generic (PLEG): container finished" podID="80b0e746-0802-4112-b77b-4db85dc85a9c" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" exitCode=143 Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.858979 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859606 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859095 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859723 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.859975 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860041 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860104 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860164 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860220 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860266 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860323 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860391 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860442 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860492 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860539 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860583 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860654 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860726 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860780 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860832 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.860949 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861033 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861122 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861193 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861250 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861300 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861353 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861403 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861453 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861507 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861564 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861613 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861716 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861798 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861860 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861932 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.861996 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862154 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862219 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862290 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lfwr6" event={"ID":"80b0e746-0802-4112-b77b-4db85dc85a9c","Type":"ContainerDied","Data":"eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862410 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862477 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862531 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862580 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862643 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862718 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862767 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862816 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862860 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862905 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.862954 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"9bcc60db49071fcce8d352c20b840bd05b05e38431a525abfef3989ca34efbd7"} Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.914531 4750 scope.go:117] "RemoveContainer" containerID="1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.920950 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lfwr6"] Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.924473 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lfwr6"] Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.948897 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:39:19 crc kubenswrapper[4750]: I0309 18:39:19.998186 4750 scope.go:117] "RemoveContainer" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.024985 4750 scope.go:117] "RemoveContainer" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.070050 4750 scope.go:117] "RemoveContainer" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.081497 4750 scope.go:117] "RemoveContainer" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.091114 4750 scope.go:117] "RemoveContainer" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.098738 4750 scope.go:117] "RemoveContainer" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.107829 4750 scope.go:117] "RemoveContainer" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.119456 4750 scope.go:117] "RemoveContainer" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.120236 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": container with ID starting with f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84 not found: ID does not exist" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.120275 4750 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": rpc error: code = NotFound desc = could not find container \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": container with ID starting with f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84 not found: ID does not exist" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.120299 4750 scope.go:117] "RemoveContainer" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.120684 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": container with ID starting with f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16 not found: ID does not exist" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.120745 4750 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": rpc error: code = NotFound desc = could not find container \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": container with ID starting with f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16 not found: ID does not exist" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.120781 4750 scope.go:117] "RemoveContainer" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.121106 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": container with ID starting with 96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163 not found: ID does not exist" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.121137 4750 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": rpc error: code = NotFound desc = could not find container \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": container with ID starting with 96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163 not found: ID does not exist" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.121158 4750 scope.go:117] "RemoveContainer" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.137087 4750 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-ovn-metrics_ovnkube-node-lfwr6_openshift-ovn-kubernetes_80b0e746-0802-4112-b77b-4db85dc85a9c_0 in pod sandbox eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93 from index: no such id: '524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9'" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.137159 4750 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-ovn-metrics_ovnkube-node-lfwr6_openshift-ovn-kubernetes_80b0e746-0802-4112-b77b-4db85dc85a9c_0 in pod sandbox eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93 from index: no such id: '524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9'" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.137167 4750 scope.go:117] "RemoveContainer" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.137300 4750 scope.go:117] "RemoveContainer" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.137897 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": container with ID starting with a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5 not found: ID does not exist" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.137954 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} err="failed to get container status \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": rpc error: code = NotFound desc = could not find container \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": container with ID starting with a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.137986 4750 scope.go:117] "RemoveContainer" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.138511 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": container with ID starting with 4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3 not found: ID does not exist" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.138574 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} err="failed to get container status \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": rpc error: code = NotFound desc = could not find container \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": container with ID starting with 4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.138613 4750 scope.go:117] "RemoveContainer" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.154831 4750 scope.go:117] "RemoveContainer" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.154827 4750 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_ovn-controller_ovnkube-node-lfwr6_openshift-ovn-kubernetes_80b0e746-0802-4112-b77b-4db85dc85a9c_0 in pod sandbox eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93 from index: no such id: '1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac'" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.155008 4750 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_ovn-controller_ovnkube-node-lfwr6_openshift-ovn-kubernetes_80b0e746-0802-4112-b77b-4db85dc85a9c_0 in pod sandbox eff357f997877a625209d19eb237bfc461fd5657e4c1183be272f6344a1c8d93 from index: no such id: '1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac'" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.155591 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": container with ID starting with 4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc not found: ID does not exist" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.155652 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} err="failed to get container status \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": rpc error: code = NotFound desc = could not find container \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": container with ID starting with 4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.155693 4750 scope.go:117] "RemoveContainer" containerID="1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.156038 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": container with ID starting with 1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae not found: ID does not exist" containerID="1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156074 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} err="failed to get container status \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": rpc error: code = NotFound desc = could not find container \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": container with ID starting with 1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156108 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.156373 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": container with ID starting with 14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6 not found: ID does not exist" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156406 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} err="failed to get container status \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": rpc error: code = NotFound desc = could not find container \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": container with ID starting with 14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156429 4750 scope.go:117] "RemoveContainer" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156683 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} err="failed to get container status \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": rpc error: code = NotFound desc = could not find container \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": container with ID starting with f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156712 4750 scope.go:117] "RemoveContainer" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156924 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} err="failed to get container status \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": rpc error: code = NotFound desc = could not find container \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": container with ID starting with 96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.156949 4750 scope.go:117] "RemoveContainer" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.157132 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} err="failed to get container status \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": rpc error: code = NotFound desc = could not find container \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": container with ID starting with f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.157158 4750 scope.go:117] "RemoveContainer" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.157362 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": container with ID starting with 524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9 not found: ID does not exist" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.157395 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} err="failed to get container status \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": rpc error: code = NotFound desc = could not find container \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": container with ID starting with 524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.157411 4750 scope.go:117] "RemoveContainer" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.157773 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} err="failed to get container status \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": rpc error: code = NotFound desc = could not find container \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": container with ID starting with a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.157804 4750 scope.go:117] "RemoveContainer" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.158028 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} err="failed to get container status \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": rpc error: code = NotFound desc = could not find container \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": container with ID starting with 4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.158052 4750 scope.go:117] "RemoveContainer" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: E0309 18:39:20.158348 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": container with ID starting with 1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac not found: ID does not exist" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.158373 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} err="failed to get container status \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": rpc error: code = NotFound desc = could not find container \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": container with ID starting with 1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.158390 4750 scope.go:117] "RemoveContainer" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.158622 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} err="failed to get container status \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": rpc error: code = NotFound desc = could not find container \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": container with ID starting with 4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.158655 4750 scope.go:117] "RemoveContainer" containerID="1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.159656 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} err="failed to get container status \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": rpc error: code = NotFound desc = could not find container \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": container with ID starting with 1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.159688 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.160771 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} err="failed to get container status \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": rpc error: code = NotFound desc = could not find container \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": container with ID starting with 14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.160803 4750 scope.go:117] "RemoveContainer" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.161358 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} err="failed to get container status \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": rpc error: code = NotFound desc = could not find container \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": container with ID starting with f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.161382 4750 scope.go:117] "RemoveContainer" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.161851 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} err="failed to get container status \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": rpc error: code = NotFound desc = could not find container \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": container with ID starting with 96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.161870 4750 scope.go:117] "RemoveContainer" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.162128 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} err="failed to get container status \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": rpc error: code = NotFound desc = could not find container \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": container with ID starting with f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.162152 4750 scope.go:117] "RemoveContainer" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.162408 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} err="failed to get container status \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": rpc error: code = NotFound desc = could not find container \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": container with ID starting with 524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.162425 4750 scope.go:117] "RemoveContainer" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.162706 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} err="failed to get container status \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": rpc error: code = NotFound desc = could not find container \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": container with ID starting with a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.162726 4750 scope.go:117] "RemoveContainer" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.162998 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} err="failed to get container status \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": rpc error: code = NotFound desc = could not find container \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": container with ID starting with 4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.163029 4750 scope.go:117] "RemoveContainer" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.163352 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} err="failed to get container status \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": rpc error: code = NotFound desc = could not find container \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": container with ID starting with 1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.163403 4750 scope.go:117] "RemoveContainer" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.163726 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} err="failed to get container status \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": rpc error: code = NotFound desc = could not find container \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": container with ID starting with 4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.163753 4750 scope.go:117] "RemoveContainer" containerID="1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.164077 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} err="failed to get container status \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": rpc error: code = NotFound desc = could not find container \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": container with ID starting with 1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.164107 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.164376 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} err="failed to get container status \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": rpc error: code = NotFound desc = could not find container \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": container with ID starting with 14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.164404 4750 scope.go:117] "RemoveContainer" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.164778 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} err="failed to get container status \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": rpc error: code = NotFound desc = could not find container \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": container with ID starting with f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.164808 4750 scope.go:117] "RemoveContainer" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165081 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} err="failed to get container status \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": rpc error: code = NotFound desc = could not find container \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": container with ID starting with 96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165107 4750 scope.go:117] "RemoveContainer" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165357 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} err="failed to get container status \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": rpc error: code = NotFound desc = could not find container \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": container with ID starting with f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165377 4750 scope.go:117] "RemoveContainer" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165660 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} err="failed to get container status \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": rpc error: code = NotFound desc = could not find container \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": container with ID starting with 524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165687 4750 scope.go:117] "RemoveContainer" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165934 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} err="failed to get container status \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": rpc error: code = NotFound desc = could not find container \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": container with ID starting with a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.165957 4750 scope.go:117] "RemoveContainer" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166186 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} err="failed to get container status \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": rpc error: code = NotFound desc = could not find container \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": container with ID starting with 4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166211 4750 scope.go:117] "RemoveContainer" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166480 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} err="failed to get container status \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": rpc error: code = NotFound desc = could not find container \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": container with ID starting with 1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166501 4750 scope.go:117] "RemoveContainer" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166710 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} err="failed to get container status \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": rpc error: code = NotFound desc = could not find container \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": container with ID starting with 4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166727 4750 scope.go:117] "RemoveContainer" containerID="1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166932 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae"} err="failed to get container status \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": rpc error: code = NotFound desc = could not find container \"1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae\": container with ID starting with 1d2e652d0c5b34ccc0af2a0a60a4c87c537774799fd63e135c85d50e41af53ae not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.166952 4750 scope.go:117] "RemoveContainer" containerID="14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167174 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6"} err="failed to get container status \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": rpc error: code = NotFound desc = could not find container \"14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6\": container with ID starting with 14057de6d777faa1fdf1266d6f2bf0a2c7e8db9b71b7882118a89fa275ad47c6 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167191 4750 scope.go:117] "RemoveContainer" containerID="f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167384 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16"} err="failed to get container status \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": rpc error: code = NotFound desc = could not find container \"f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16\": container with ID starting with f0318b3a0092b1f8290a7451f4c6a0db1ced1f0af809ff68ad6876decc259a16 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167401 4750 scope.go:117] "RemoveContainer" containerID="96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167577 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163"} err="failed to get container status \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": rpc error: code = NotFound desc = could not find container \"96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163\": container with ID starting with 96adc59b508c5c3c090fbf9df2d47036648e9f08567fde1de8d562d899635163 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167595 4750 scope.go:117] "RemoveContainer" containerID="f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167805 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84"} err="failed to get container status \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": rpc error: code = NotFound desc = could not find container \"f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84\": container with ID starting with f20222840139d0d1fcfea688bae5505e543a2b2a7bd0d0f24fc15e1810c2ea84 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167822 4750 scope.go:117] "RemoveContainer" containerID="524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.167998 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9"} err="failed to get container status \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": rpc error: code = NotFound desc = could not find container \"524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9\": container with ID starting with 524ae95f56e60e71a150be01cdc1ebf1fa684c8e0cc4d2788677c7bfe13576d9 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168013 4750 scope.go:117] "RemoveContainer" containerID="a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168183 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5"} err="failed to get container status \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": rpc error: code = NotFound desc = could not find container \"a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5\": container with ID starting with a2911363afd697c94fe329423e4b73c1224567d3899f190a905f3c2aa985a8e5 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168202 4750 scope.go:117] "RemoveContainer" containerID="4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168367 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3"} err="failed to get container status \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": rpc error: code = NotFound desc = could not find container \"4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3\": container with ID starting with 4256d8ca8f1ebc97623319b756cb33e00f6867ca94a66220b6edc9c1ecb38cb3 not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168387 4750 scope.go:117] "RemoveContainer" containerID="1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168574 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac"} err="failed to get container status \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": rpc error: code = NotFound desc = could not find container \"1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac\": container with ID starting with 1437ad1e404617e702baf00f00ca3bfbaf24598d590412c54e5ab85330e5afac not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168590 4750 scope.go:117] "RemoveContainer" containerID="4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.168787 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc"} err="failed to get container status \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": rpc error: code = NotFound desc = could not find container \"4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc\": container with ID starting with 4fb47d1ca84c954cfdd0a12db1d36e70510bfc0c7feb13ef59973111220236bc not found: ID does not exist" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.879816 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vj6kf_fa652827-92a6-4544-b1b6-cba227f8e9f4/kube-multus/2.log" Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.880017 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vj6kf" event={"ID":"fa652827-92a6-4544-b1b6-cba227f8e9f4","Type":"ContainerStarted","Data":"848bccb8470eef430e7734c63287d3ae9efcad5b4dcb1c923db8471c4c48fba2"} Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.883378 4750 generic.go:334] "Generic (PLEG): container finished" podID="ee99c41c-bd53-4dd7-be2b-a8916902a942" containerID="312977cba816e54fd5eaf8548b56024cc56a0aa391556b189f5a8e481e8c1b51" exitCode=0 Mar 09 18:39:20 crc kubenswrapper[4750]: I0309 18:39:20.883467 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerDied","Data":"312977cba816e54fd5eaf8548b56024cc56a0aa391556b189f5a8e481e8c1b51"} Mar 09 18:39:21 crc kubenswrapper[4750]: I0309 18:39:21.382826 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b0e746-0802-4112-b77b-4db85dc85a9c" path="/var/lib/kubelet/pods/80b0e746-0802-4112-b77b-4db85dc85a9c/volumes" Mar 09 18:39:21 crc kubenswrapper[4750]: I0309 18:39:21.894873 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"27684de0331a87c66346388be4aeab912692fde4d489f08aba52d5408c74c0ef"} Mar 09 18:39:21 crc kubenswrapper[4750]: I0309 18:39:21.894952 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"77b3440c89d0bf5e5f678dcdb6b269083e6e145b94def9f2f8af90fe6c40959c"} Mar 09 18:39:21 crc kubenswrapper[4750]: I0309 18:39:21.894974 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"ae93cf928a19ece3f8199fe189788579bde7460adfc033119bb8545ea2c7ce1e"} Mar 09 18:39:21 crc kubenswrapper[4750]: I0309 18:39:21.894994 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"d4a95bdf30766fa5afac5fc9e3d3749e43b1973d992e92a6ac83790fb7ffe164"} Mar 09 18:39:21 crc kubenswrapper[4750]: I0309 18:39:21.895013 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"0acc8405d21fbb8ec950ea760bbe7f9d11c6005e4819d2317629fe6fa8002b9b"} Mar 09 18:39:21 crc kubenswrapper[4750]: I0309 18:39:21.895033 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"20c3936184907c3848ad220c8189fb3726d3676a2066220c31e5364ed62e1c71"} Mar 09 18:39:24 crc kubenswrapper[4750]: I0309 18:39:24.918401 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"f60d296add2f1dbe500ea4c2ec577fcaae98a6281a12a0dc9f4e9c3cf9b9490a"} Mar 09 18:39:27 crc kubenswrapper[4750]: I0309 18:39:27.941949 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" event={"ID":"ee99c41c-bd53-4dd7-be2b-a8916902a942","Type":"ContainerStarted","Data":"21b031b0c64ece8f52f4af0779fff928ca697e40e7c77facbcd4452f9c4f2f3b"} Mar 09 18:39:27 crc kubenswrapper[4750]: I0309 18:39:27.942827 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:27 crc kubenswrapper[4750]: I0309 18:39:27.942848 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:27 crc kubenswrapper[4750]: I0309 18:39:27.942863 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:27 crc kubenswrapper[4750]: I0309 18:39:27.983252 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" podStartSLOduration=8.983232895 podStartE2EDuration="8.983232895s" podCreationTimestamp="2026-03-09 18:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:39:27.978193321 +0000 UTC m=+849.320665719" watchObservedRunningTime="2026-03-09 18:39:27.983232895 +0000 UTC m=+849.325705313" Mar 09 18:39:28 crc kubenswrapper[4750]: I0309 18:39:28.036267 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:28 crc kubenswrapper[4750]: I0309 18:39:28.040317 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:36 crc kubenswrapper[4750]: I0309 18:39:36.298912 4750 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.433851 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6"] Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.437163 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.439595 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.454603 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6"] Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.566472 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.566568 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.566687 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmqnt\" (UniqueName: \"kubernetes.io/projected/908909ba-f493-489e-834e-863230ee979b-kube-api-access-jmqnt\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.667772 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.667836 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.667871 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmqnt\" (UniqueName: \"kubernetes.io/projected/908909ba-f493-489e-834e-863230ee979b-kube-api-access-jmqnt\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.668577 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.668995 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.693074 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmqnt\" (UniqueName: \"kubernetes.io/projected/908909ba-f493-489e-834e-863230ee979b-kube-api-access-jmqnt\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:47 crc kubenswrapper[4750]: I0309 18:39:47.768466 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:48 crc kubenswrapper[4750]: I0309 18:39:48.048173 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6"] Mar 09 18:39:48 crc kubenswrapper[4750]: W0309 18:39:48.055543 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod908909ba_f493_489e_834e_863230ee979b.slice/crio-3a26605cf9b496964f438c493f9357ef9fc98be57784cb9c65a4852cbbc04f9a WatchSource:0}: Error finding container 3a26605cf9b496964f438c493f9357ef9fc98be57784cb9c65a4852cbbc04f9a: Status 404 returned error can't find the container with id 3a26605cf9b496964f438c493f9357ef9fc98be57784cb9c65a4852cbbc04f9a Mar 09 18:39:48 crc kubenswrapper[4750]: I0309 18:39:48.082327 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" event={"ID":"908909ba-f493-489e-834e-863230ee979b","Type":"ContainerStarted","Data":"3a26605cf9b496964f438c493f9357ef9fc98be57784cb9c65a4852cbbc04f9a"} Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.091910 4750 generic.go:334] "Generic (PLEG): container finished" podID="908909ba-f493-489e-834e-863230ee979b" containerID="b9c37813b7b776c3c08e614abc15f91b69b211e3c1f89a95a85d12bffaca5015" exitCode=0 Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.091982 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" event={"ID":"908909ba-f493-489e-834e-863230ee979b","Type":"ContainerDied","Data":"b9c37813b7b776c3c08e614abc15f91b69b211e3c1f89a95a85d12bffaca5015"} Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.757808 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7xpkz"] Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.760067 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.776782 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xpkz"] Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.809043 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hx78v" Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.900621 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8tnc\" (UniqueName: \"kubernetes.io/projected/29de4cb5-a1ae-446a-8e27-195aecca6b1a-kube-api-access-n8tnc\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.900732 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-utilities\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:49 crc kubenswrapper[4750]: I0309 18:39:49.900795 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-catalog-content\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.002580 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-utilities\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.003067 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-catalog-content\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.003769 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8tnc\" (UniqueName: \"kubernetes.io/projected/29de4cb5-a1ae-446a-8e27-195aecca6b1a-kube-api-access-n8tnc\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.003662 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-catalog-content\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.003217 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-utilities\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.028539 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8tnc\" (UniqueName: \"kubernetes.io/projected/29de4cb5-a1ae-446a-8e27-195aecca6b1a-kube-api-access-n8tnc\") pod \"redhat-operators-7xpkz\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.096792 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:39:50 crc kubenswrapper[4750]: I0309 18:39:50.612940 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xpkz"] Mar 09 18:39:51 crc kubenswrapper[4750]: I0309 18:39:51.110366 4750 generic.go:334] "Generic (PLEG): container finished" podID="908909ba-f493-489e-834e-863230ee979b" containerID="e8a090766f1e975f3f591232a8aa349149fd3be0426c5637af4c73eaf7af7f39" exitCode=0 Mar 09 18:39:51 crc kubenswrapper[4750]: I0309 18:39:51.110422 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" event={"ID":"908909ba-f493-489e-834e-863230ee979b","Type":"ContainerDied","Data":"e8a090766f1e975f3f591232a8aa349149fd3be0426c5637af4c73eaf7af7f39"} Mar 09 18:39:51 crc kubenswrapper[4750]: I0309 18:39:51.114291 4750 generic.go:334] "Generic (PLEG): container finished" podID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerID="c4febc759a2519cf2162d1059d2a09b6fd9f37967ce8d19d6f7cecd9e7799647" exitCode=0 Mar 09 18:39:51 crc kubenswrapper[4750]: I0309 18:39:51.114405 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xpkz" event={"ID":"29de4cb5-a1ae-446a-8e27-195aecca6b1a","Type":"ContainerDied","Data":"c4febc759a2519cf2162d1059d2a09b6fd9f37967ce8d19d6f7cecd9e7799647"} Mar 09 18:39:51 crc kubenswrapper[4750]: I0309 18:39:51.114456 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xpkz" event={"ID":"29de4cb5-a1ae-446a-8e27-195aecca6b1a","Type":"ContainerStarted","Data":"52d76e8867918c262fdf74ec85fe809b71938678043559e2c8c8d9b1adf7b69c"} Mar 09 18:39:51 crc kubenswrapper[4750]: I0309 18:39:51.743217 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:39:51 crc kubenswrapper[4750]: I0309 18:39:51.743324 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:39:52 crc kubenswrapper[4750]: I0309 18:39:52.123600 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xpkz" event={"ID":"29de4cb5-a1ae-446a-8e27-195aecca6b1a","Type":"ContainerStarted","Data":"5ecc6fd4e05e78d1030d71ac017570c7d2c424c6d21c7a64ae1523d2bbae86fb"} Mar 09 18:39:52 crc kubenswrapper[4750]: I0309 18:39:52.127942 4750 generic.go:334] "Generic (PLEG): container finished" podID="908909ba-f493-489e-834e-863230ee979b" containerID="0279f891f31544b51e4b5afeb77c7750e4a2e1f9aae2860f63ade40fc46eb9d6" exitCode=0 Mar 09 18:39:52 crc kubenswrapper[4750]: I0309 18:39:52.127979 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" event={"ID":"908909ba-f493-489e-834e-863230ee979b","Type":"ContainerDied","Data":"0279f891f31544b51e4b5afeb77c7750e4a2e1f9aae2860f63ade40fc46eb9d6"} Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.141746 4750 generic.go:334] "Generic (PLEG): container finished" podID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerID="5ecc6fd4e05e78d1030d71ac017570c7d2c424c6d21c7a64ae1523d2bbae86fb" exitCode=0 Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.141922 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xpkz" event={"ID":"29de4cb5-a1ae-446a-8e27-195aecca6b1a","Type":"ContainerDied","Data":"5ecc6fd4e05e78d1030d71ac017570c7d2c424c6d21c7a64ae1523d2bbae86fb"} Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.549208 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.658411 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-bundle\") pod \"908909ba-f493-489e-834e-863230ee979b\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.658501 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-util\") pod \"908909ba-f493-489e-834e-863230ee979b\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.658551 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmqnt\" (UniqueName: \"kubernetes.io/projected/908909ba-f493-489e-834e-863230ee979b-kube-api-access-jmqnt\") pod \"908909ba-f493-489e-834e-863230ee979b\" (UID: \"908909ba-f493-489e-834e-863230ee979b\") " Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.662804 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-bundle" (OuterVolumeSpecName: "bundle") pod "908909ba-f493-489e-834e-863230ee979b" (UID: "908909ba-f493-489e-834e-863230ee979b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.679148 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/908909ba-f493-489e-834e-863230ee979b-kube-api-access-jmqnt" (OuterVolumeSpecName: "kube-api-access-jmqnt") pod "908909ba-f493-489e-834e-863230ee979b" (UID: "908909ba-f493-489e-834e-863230ee979b"). InnerVolumeSpecName "kube-api-access-jmqnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.704994 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-util" (OuterVolumeSpecName: "util") pod "908909ba-f493-489e-834e-863230ee979b" (UID: "908909ba-f493-489e-834e-863230ee979b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.760429 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.760465 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/908909ba-f493-489e-834e-863230ee979b-util\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:53 crc kubenswrapper[4750]: I0309 18:39:53.760477 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmqnt\" (UniqueName: \"kubernetes.io/projected/908909ba-f493-489e-834e-863230ee979b-kube-api-access-jmqnt\") on node \"crc\" DevicePath \"\"" Mar 09 18:39:54 crc kubenswrapper[4750]: I0309 18:39:54.150307 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" event={"ID":"908909ba-f493-489e-834e-863230ee979b","Type":"ContainerDied","Data":"3a26605cf9b496964f438c493f9357ef9fc98be57784cb9c65a4852cbbc04f9a"} Mar 09 18:39:54 crc kubenswrapper[4750]: I0309 18:39:54.150356 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a26605cf9b496964f438c493f9357ef9fc98be57784cb9c65a4852cbbc04f9a" Mar 09 18:39:54 crc kubenswrapper[4750]: I0309 18:39:54.150433 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6" Mar 09 18:39:54 crc kubenswrapper[4750]: I0309 18:39:54.153299 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xpkz" event={"ID":"29de4cb5-a1ae-446a-8e27-195aecca6b1a","Type":"ContainerStarted","Data":"da95bb9ddf93dfdbff6ce46d6e4d94a50ce7850c9246d1a3c2ae17b944a6a6b9"} Mar 09 18:39:54 crc kubenswrapper[4750]: I0309 18:39:54.602093 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7xpkz" podStartSLOduration=3.060999063 podStartE2EDuration="5.602060927s" podCreationTimestamp="2026-03-09 18:39:49 +0000 UTC" firstStartedPulling="2026-03-09 18:39:51.116874412 +0000 UTC m=+872.459346820" lastFinishedPulling="2026-03-09 18:39:53.657936256 +0000 UTC m=+875.000408684" observedRunningTime="2026-03-09 18:39:54.181229306 +0000 UTC m=+875.523701714" watchObservedRunningTime="2026-03-09 18:39:54.602060927 +0000 UTC m=+875.944533365" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.097926 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.098676 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.193062 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551360-plcvg"] Mar 09 18:40:00 crc kubenswrapper[4750]: E0309 18:40:00.193325 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="908909ba-f493-489e-834e-863230ee979b" containerName="extract" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.193339 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="908909ba-f493-489e-834e-863230ee979b" containerName="extract" Mar 09 18:40:00 crc kubenswrapper[4750]: E0309 18:40:00.193359 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="908909ba-f493-489e-834e-863230ee979b" containerName="pull" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.193365 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="908909ba-f493-489e-834e-863230ee979b" containerName="pull" Mar 09 18:40:00 crc kubenswrapper[4750]: E0309 18:40:00.193374 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="908909ba-f493-489e-834e-863230ee979b" containerName="util" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.193380 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="908909ba-f493-489e-834e-863230ee979b" containerName="util" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.193492 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="908909ba-f493-489e-834e-863230ee979b" containerName="extract" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.193907 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551360-plcvg" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.196999 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.197014 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.198408 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.205362 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551360-plcvg"] Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.251580 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbd7b\" (UniqueName: \"kubernetes.io/projected/5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63-kube-api-access-qbd7b\") pod \"auto-csr-approver-29551360-plcvg\" (UID: \"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63\") " pod="openshift-infra/auto-csr-approver-29551360-plcvg" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.353278 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbd7b\" (UniqueName: \"kubernetes.io/projected/5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63-kube-api-access-qbd7b\") pod \"auto-csr-approver-29551360-plcvg\" (UID: \"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63\") " pod="openshift-infra/auto-csr-approver-29551360-plcvg" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.378550 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbd7b\" (UniqueName: \"kubernetes.io/projected/5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63-kube-api-access-qbd7b\") pod \"auto-csr-approver-29551360-plcvg\" (UID: \"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63\") " pod="openshift-infra/auto-csr-approver-29551360-plcvg" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.510599 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551360-plcvg" Mar 09 18:40:00 crc kubenswrapper[4750]: I0309 18:40:00.770936 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551360-plcvg"] Mar 09 18:40:01 crc kubenswrapper[4750]: I0309 18:40:01.143456 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7xpkz" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="registry-server" probeResult="failure" output=< Mar 09 18:40:01 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 18:40:01 crc kubenswrapper[4750]: > Mar 09 18:40:01 crc kubenswrapper[4750]: I0309 18:40:01.228619 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551360-plcvg" event={"ID":"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63","Type":"ContainerStarted","Data":"80e1917b52316518787d866f778a5fa443e07fdd91cba0503e6eb10bb9be9d23"} Mar 09 18:40:02 crc kubenswrapper[4750]: I0309 18:40:02.243238 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551360-plcvg" event={"ID":"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63","Type":"ContainerStarted","Data":"b54aed15db155e225dfc62a44b1737b87c9e40783d39c0fdc434ca62f0803160"} Mar 09 18:40:02 crc kubenswrapper[4750]: I0309 18:40:02.263979 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551360-plcvg" podStartSLOduration=1.261536135 podStartE2EDuration="2.263950649s" podCreationTimestamp="2026-03-09 18:40:00 +0000 UTC" firstStartedPulling="2026-03-09 18:40:00.799613432 +0000 UTC m=+882.142085830" lastFinishedPulling="2026-03-09 18:40:01.802027936 +0000 UTC m=+883.144500344" observedRunningTime="2026-03-09 18:40:02.262091919 +0000 UTC m=+883.604564317" watchObservedRunningTime="2026-03-09 18:40:02.263950649 +0000 UTC m=+883.606423047" Mar 09 18:40:04 crc kubenswrapper[4750]: I0309 18:40:04.255218 4750 generic.go:334] "Generic (PLEG): container finished" podID="5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63" containerID="b54aed15db155e225dfc62a44b1737b87c9e40783d39c0fdc434ca62f0803160" exitCode=0 Mar 09 18:40:04 crc kubenswrapper[4750]: I0309 18:40:04.255310 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551360-plcvg" event={"ID":"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63","Type":"ContainerDied","Data":"b54aed15db155e225dfc62a44b1737b87c9e40783d39c0fdc434ca62f0803160"} Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.426583 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.433498 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.440088 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.440250 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.440361 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-hcj87" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.442747 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.489786 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.491232 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.495720 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.496043 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-n4mdw" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.509709 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.510668 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.520014 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.523313 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz\" (UID: \"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.523367 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b89xt\" (UniqueName: \"kubernetes.io/projected/8581d9a9-567b-43cc-919e-92099fcffc93-kube-api-access-b89xt\") pod \"obo-prometheus-operator-68bc856cb9-7kjhn\" (UID: \"8581d9a9-567b-43cc-919e-92099fcffc93\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.523397 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz\" (UID: \"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.523468 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.625450 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5ff4820-7371-4401-805c-862b76511e97-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt\" (UID: \"e5ff4820-7371-4401-805c-862b76511e97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.625541 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz\" (UID: \"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.625568 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b89xt\" (UniqueName: \"kubernetes.io/projected/8581d9a9-567b-43cc-919e-92099fcffc93-kube-api-access-b89xt\") pod \"obo-prometheus-operator-68bc856cb9-7kjhn\" (UID: \"8581d9a9-567b-43cc-919e-92099fcffc93\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.625597 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz\" (UID: \"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.625683 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5ff4820-7371-4401-805c-862b76511e97-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt\" (UID: \"e5ff4820-7371-4401-805c-862b76511e97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.634054 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz\" (UID: \"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.637239 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz\" (UID: \"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.646821 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551360-plcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.649365 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b89xt\" (UniqueName: \"kubernetes.io/projected/8581d9a9-567b-43cc-919e-92099fcffc93-kube-api-access-b89xt\") pod \"obo-prometheus-operator-68bc856cb9-7kjhn\" (UID: \"8581d9a9-567b-43cc-919e-92099fcffc93\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.652169 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-jbcvg"] Mar 09 18:40:05 crc kubenswrapper[4750]: E0309 18:40:05.652387 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63" containerName="oc" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.652404 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63" containerName="oc" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.652507 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63" containerName="oc" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.659164 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.663965 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-99xvs" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.664333 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.681757 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-jbcvg"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.727010 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbd7b\" (UniqueName: \"kubernetes.io/projected/5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63-kube-api-access-qbd7b\") pod \"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63\" (UID: \"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63\") " Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.727228 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5ff4820-7371-4401-805c-862b76511e97-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt\" (UID: \"e5ff4820-7371-4401-805c-862b76511e97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.727277 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llmsc\" (UniqueName: \"kubernetes.io/projected/9b0f552e-2e27-47d5-a0ca-49a041ddb74a-kube-api-access-llmsc\") pod \"observability-operator-59bdc8b94-jbcvg\" (UID: \"9b0f552e-2e27-47d5-a0ca-49a041ddb74a\") " pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.727311 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5ff4820-7371-4401-805c-862b76511e97-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt\" (UID: \"e5ff4820-7371-4401-805c-862b76511e97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.727338 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b0f552e-2e27-47d5-a0ca-49a041ddb74a-observability-operator-tls\") pod \"observability-operator-59bdc8b94-jbcvg\" (UID: \"9b0f552e-2e27-47d5-a0ca-49a041ddb74a\") " pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.731610 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63-kube-api-access-qbd7b" (OuterVolumeSpecName: "kube-api-access-qbd7b") pod "5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63" (UID: "5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63"). InnerVolumeSpecName "kube-api-access-qbd7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.734060 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5ff4820-7371-4401-805c-862b76511e97-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt\" (UID: \"e5ff4820-7371-4401-805c-862b76511e97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.749147 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5ff4820-7371-4401-805c-862b76511e97-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt\" (UID: \"e5ff4820-7371-4401-805c-862b76511e97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.788589 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.820868 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.829349 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llmsc\" (UniqueName: \"kubernetes.io/projected/9b0f552e-2e27-47d5-a0ca-49a041ddb74a-kube-api-access-llmsc\") pod \"observability-operator-59bdc8b94-jbcvg\" (UID: \"9b0f552e-2e27-47d5-a0ca-49a041ddb74a\") " pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.829468 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b0f552e-2e27-47d5-a0ca-49a041ddb74a-observability-operator-tls\") pod \"observability-operator-59bdc8b94-jbcvg\" (UID: \"9b0f552e-2e27-47d5-a0ca-49a041ddb74a\") " pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.829531 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbd7b\" (UniqueName: \"kubernetes.io/projected/5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63-kube-api-access-qbd7b\") on node \"crc\" DevicePath \"\"" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.835979 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.840307 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b0f552e-2e27-47d5-a0ca-49a041ddb74a-observability-operator-tls\") pod \"observability-operator-59bdc8b94-jbcvg\" (UID: \"9b0f552e-2e27-47d5-a0ca-49a041ddb74a\") " pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.852936 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llmsc\" (UniqueName: \"kubernetes.io/projected/9b0f552e-2e27-47d5-a0ca-49a041ddb74a-kube-api-access-llmsc\") pod \"observability-operator-59bdc8b94-jbcvg\" (UID: \"9b0f552e-2e27-47d5-a0ca-49a041ddb74a\") " pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.866326 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-7cmcp"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.867462 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.871307 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-xkhcq" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.890490 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-7cmcp"] Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.931535 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq6mj\" (UniqueName: \"kubernetes.io/projected/ef54df65-856d-4a28-a4ba-bec1bf76ac07-kube-api-access-cq6mj\") pod \"perses-operator-5bf474d74f-7cmcp\" (UID: \"ef54df65-856d-4a28-a4ba-bec1bf76ac07\") " pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.931652 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ef54df65-856d-4a28-a4ba-bec1bf76ac07-openshift-service-ca\") pod \"perses-operator-5bf474d74f-7cmcp\" (UID: \"ef54df65-856d-4a28-a4ba-bec1bf76ac07\") " pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:05 crc kubenswrapper[4750]: I0309 18:40:05.994358 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.032887 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq6mj\" (UniqueName: \"kubernetes.io/projected/ef54df65-856d-4a28-a4ba-bec1bf76ac07-kube-api-access-cq6mj\") pod \"perses-operator-5bf474d74f-7cmcp\" (UID: \"ef54df65-856d-4a28-a4ba-bec1bf76ac07\") " pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.032984 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ef54df65-856d-4a28-a4ba-bec1bf76ac07-openshift-service-ca\") pod \"perses-operator-5bf474d74f-7cmcp\" (UID: \"ef54df65-856d-4a28-a4ba-bec1bf76ac07\") " pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.036692 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ef54df65-856d-4a28-a4ba-bec1bf76ac07-openshift-service-ca\") pod \"perses-operator-5bf474d74f-7cmcp\" (UID: \"ef54df65-856d-4a28-a4ba-bec1bf76ac07\") " pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.098529 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq6mj\" (UniqueName: \"kubernetes.io/projected/ef54df65-856d-4a28-a4ba-bec1bf76ac07-kube-api-access-cq6mj\") pod \"perses-operator-5bf474d74f-7cmcp\" (UID: \"ef54df65-856d-4a28-a4ba-bec1bf76ac07\") " pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.192088 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.238592 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz"] Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.283022 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551360-plcvg" event={"ID":"5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63","Type":"ContainerDied","Data":"80e1917b52316518787d866f778a5fa443e07fdd91cba0503e6eb10bb9be9d23"} Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.283058 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551360-plcvg" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.283082 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80e1917b52316518787d866f778a5fa443e07fdd91cba0503e6eb10bb9be9d23" Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.293352 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn"] Mar 09 18:40:06 crc kubenswrapper[4750]: W0309 18:40:06.307149 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8581d9a9_567b_43cc_919e_92099fcffc93.slice/crio-718f8c7f47809d092a563add8ee05613330dd903d6e96fcf86dd52783fd453ef WatchSource:0}: Error finding container 718f8c7f47809d092a563add8ee05613330dd903d6e96fcf86dd52783fd453ef: Status 404 returned error can't find the container with id 718f8c7f47809d092a563add8ee05613330dd903d6e96fcf86dd52783fd453ef Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.404118 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt"] Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.560168 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-7cmcp"] Mar 09 18:40:06 crc kubenswrapper[4750]: W0309 18:40:06.569994 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef54df65_856d_4a28_a4ba_bec1bf76ac07.slice/crio-b975c4bba514ebcbb19a0805ec8996ec92e455c2b94a9e74b2d502d778f4cbb7 WatchSource:0}: Error finding container b975c4bba514ebcbb19a0805ec8996ec92e455c2b94a9e74b2d502d778f4cbb7: Status 404 returned error can't find the container with id b975c4bba514ebcbb19a0805ec8996ec92e455c2b94a9e74b2d502d778f4cbb7 Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.683118 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-jbcvg"] Mar 09 18:40:06 crc kubenswrapper[4750]: W0309 18:40:06.691337 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b0f552e_2e27_47d5_a0ca_49a041ddb74a.slice/crio-606f152630eeb8e1cf7e506ed4472a141dfe4edc6c09a3f64c2d8a7c9f961be2 WatchSource:0}: Error finding container 606f152630eeb8e1cf7e506ed4472a141dfe4edc6c09a3f64c2d8a7c9f961be2: Status 404 returned error can't find the container with id 606f152630eeb8e1cf7e506ed4472a141dfe4edc6c09a3f64c2d8a7c9f961be2 Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.751488 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551354-fpht2"] Mar 09 18:40:06 crc kubenswrapper[4750]: I0309 18:40:06.760176 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551354-fpht2"] Mar 09 18:40:07 crc kubenswrapper[4750]: I0309 18:40:07.290569 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" event={"ID":"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86","Type":"ContainerStarted","Data":"56a8a03601efbef242f771432c44dfaefe54ad8e52ab063406fda07ec8cf04ba"} Mar 09 18:40:07 crc kubenswrapper[4750]: I0309 18:40:07.291515 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" event={"ID":"ef54df65-856d-4a28-a4ba-bec1bf76ac07","Type":"ContainerStarted","Data":"b975c4bba514ebcbb19a0805ec8996ec92e455c2b94a9e74b2d502d778f4cbb7"} Mar 09 18:40:07 crc kubenswrapper[4750]: I0309 18:40:07.292451 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" event={"ID":"9b0f552e-2e27-47d5-a0ca-49a041ddb74a","Type":"ContainerStarted","Data":"606f152630eeb8e1cf7e506ed4472a141dfe4edc6c09a3f64c2d8a7c9f961be2"} Mar 09 18:40:07 crc kubenswrapper[4750]: I0309 18:40:07.294033 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" event={"ID":"e5ff4820-7371-4401-805c-862b76511e97","Type":"ContainerStarted","Data":"f7b3ea2dc36ebcb8e46394d37ce03734c2a2498c151fe7d24505fdf6486a39ef"} Mar 09 18:40:07 crc kubenswrapper[4750]: I0309 18:40:07.295230 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" event={"ID":"8581d9a9-567b-43cc-919e-92099fcffc93","Type":"ContainerStarted","Data":"718f8c7f47809d092a563add8ee05613330dd903d6e96fcf86dd52783fd453ef"} Mar 09 18:40:07 crc kubenswrapper[4750]: I0309 18:40:07.380010 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c072a7f-6cb4-4857-9425-ca75503d6470" path="/var/lib/kubelet/pods/6c072a7f-6cb4-4857-9425-ca75503d6470/volumes" Mar 09 18:40:10 crc kubenswrapper[4750]: I0309 18:40:10.159534 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:40:10 crc kubenswrapper[4750]: I0309 18:40:10.240403 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:40:12 crc kubenswrapper[4750]: I0309 18:40:12.153655 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xpkz"] Mar 09 18:40:12 crc kubenswrapper[4750]: I0309 18:40:12.153951 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7xpkz" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="registry-server" containerID="cri-o://da95bb9ddf93dfdbff6ce46d6e4d94a50ce7850c9246d1a3c2ae17b944a6a6b9" gracePeriod=2 Mar 09 18:40:12 crc kubenswrapper[4750]: I0309 18:40:12.347945 4750 generic.go:334] "Generic (PLEG): container finished" podID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerID="da95bb9ddf93dfdbff6ce46d6e4d94a50ce7850c9246d1a3c2ae17b944a6a6b9" exitCode=0 Mar 09 18:40:12 crc kubenswrapper[4750]: I0309 18:40:12.348002 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xpkz" event={"ID":"29de4cb5-a1ae-446a-8e27-195aecca6b1a","Type":"ContainerDied","Data":"da95bb9ddf93dfdbff6ce46d6e4d94a50ce7850c9246d1a3c2ae17b944a6a6b9"} Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.283827 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.389022 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xpkz" event={"ID":"29de4cb5-a1ae-446a-8e27-195aecca6b1a","Type":"ContainerDied","Data":"52d76e8867918c262fdf74ec85fe809b71938678043559e2c8c8d9b1adf7b69c"} Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.389085 4750 scope.go:117] "RemoveContainer" containerID="da95bb9ddf93dfdbff6ce46d6e4d94a50ce7850c9246d1a3c2ae17b944a6a6b9" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.389236 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xpkz" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.435228 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-catalog-content\") pod \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.435280 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-utilities\") pod \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.435344 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8tnc\" (UniqueName: \"kubernetes.io/projected/29de4cb5-a1ae-446a-8e27-195aecca6b1a-kube-api-access-n8tnc\") pod \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\" (UID: \"29de4cb5-a1ae-446a-8e27-195aecca6b1a\") " Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.436703 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-utilities" (OuterVolumeSpecName: "utilities") pod "29de4cb5-a1ae-446a-8e27-195aecca6b1a" (UID: "29de4cb5-a1ae-446a-8e27-195aecca6b1a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.442318 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29de4cb5-a1ae-446a-8e27-195aecca6b1a-kube-api-access-n8tnc" (OuterVolumeSpecName: "kube-api-access-n8tnc") pod "29de4cb5-a1ae-446a-8e27-195aecca6b1a" (UID: "29de4cb5-a1ae-446a-8e27-195aecca6b1a"). InnerVolumeSpecName "kube-api-access-n8tnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.442850 4750 scope.go:117] "RemoveContainer" containerID="5ecc6fd4e05e78d1030d71ac017570c7d2c424c6d21c7a64ae1523d2bbae86fb" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.467554 4750 scope.go:117] "RemoveContainer" containerID="c4febc759a2519cf2162d1059d2a09b6fd9f37967ce8d19d6f7cecd9e7799647" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.536670 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.536722 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8tnc\" (UniqueName: \"kubernetes.io/projected/29de4cb5-a1ae-446a-8e27-195aecca6b1a-kube-api-access-n8tnc\") on node \"crc\" DevicePath \"\"" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.568619 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29de4cb5-a1ae-446a-8e27-195aecca6b1a" (UID: "29de4cb5-a1ae-446a-8e27-195aecca6b1a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.638859 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29de4cb5-a1ae-446a-8e27-195aecca6b1a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.724829 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xpkz"] Mar 09 18:40:17 crc kubenswrapper[4750]: I0309 18:40:17.730346 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7xpkz"] Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.398286 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" event={"ID":"8581d9a9-567b-43cc-919e-92099fcffc93","Type":"ContainerStarted","Data":"931c590cd3007b6456df6f7fca5db28d20a116ea674790758494f9034b38c4e3"} Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.400075 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" event={"ID":"f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86","Type":"ContainerStarted","Data":"6814dd1b7edfe174b8d93c35235ffcff53b66e8e73f869dcf3f44ef2c3beb35c"} Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.401834 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" event={"ID":"ef54df65-856d-4a28-a4ba-bec1bf76ac07","Type":"ContainerStarted","Data":"20dc42473aa17d74886072b4a5f73f1bfdc6ce711e64a84e890bdf2d17551f57"} Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.402701 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.405950 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" event={"ID":"9b0f552e-2e27-47d5-a0ca-49a041ddb74a","Type":"ContainerStarted","Data":"12ac3df4bc222167bf768ed6a951ec3a48604c369b1fd428169684777ba1d46e"} Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.406297 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.408533 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" event={"ID":"e5ff4820-7371-4401-805c-862b76511e97","Type":"ContainerStarted","Data":"8befbba3fa7a8e9ca5852d1e8ef17101db35fb8b2d50b028766cdae10f5341eb"} Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.409381 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.424422 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7kjhn" podStartSLOduration=2.479880497 podStartE2EDuration="13.424399659s" podCreationTimestamp="2026-03-09 18:40:05 +0000 UTC" firstStartedPulling="2026-03-09 18:40:06.314511022 +0000 UTC m=+887.656983420" lastFinishedPulling="2026-03-09 18:40:17.259030184 +0000 UTC m=+898.601502582" observedRunningTime="2026-03-09 18:40:18.421697907 +0000 UTC m=+899.764170305" watchObservedRunningTime="2026-03-09 18:40:18.424399659 +0000 UTC m=+899.766872057" Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.439113 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" podStartSLOduration=2.790562361 podStartE2EDuration="13.439093643s" podCreationTimestamp="2026-03-09 18:40:05 +0000 UTC" firstStartedPulling="2026-03-09 18:40:06.572690088 +0000 UTC m=+887.915162486" lastFinishedPulling="2026-03-09 18:40:17.22122137 +0000 UTC m=+898.563693768" observedRunningTime="2026-03-09 18:40:18.437872621 +0000 UTC m=+899.780345019" watchObservedRunningTime="2026-03-09 18:40:18.439093643 +0000 UTC m=+899.781566041" Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.458714 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz" podStartSLOduration=2.548380145 podStartE2EDuration="13.458682839s" podCreationTimestamp="2026-03-09 18:40:05 +0000 UTC" firstStartedPulling="2026-03-09 18:40:06.307993807 +0000 UTC m=+887.650466205" lastFinishedPulling="2026-03-09 18:40:17.218296501 +0000 UTC m=+898.560768899" observedRunningTime="2026-03-09 18:40:18.456424129 +0000 UTC m=+899.798896527" watchObservedRunningTime="2026-03-09 18:40:18.458682839 +0000 UTC m=+899.801155237" Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.485781 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-jbcvg" podStartSLOduration=2.922767621 podStartE2EDuration="13.485752996s" podCreationTimestamp="2026-03-09 18:40:05 +0000 UTC" firstStartedPulling="2026-03-09 18:40:06.693719036 +0000 UTC m=+888.036191434" lastFinishedPulling="2026-03-09 18:40:17.256704421 +0000 UTC m=+898.599176809" observedRunningTime="2026-03-09 18:40:18.478729117 +0000 UTC m=+899.821201515" watchObservedRunningTime="2026-03-09 18:40:18.485752996 +0000 UTC m=+899.828225394" Mar 09 18:40:18 crc kubenswrapper[4750]: I0309 18:40:18.523034 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt" podStartSLOduration=2.684877657 podStartE2EDuration="13.523013245s" podCreationTimestamp="2026-03-09 18:40:05 +0000 UTC" firstStartedPulling="2026-03-09 18:40:06.419587181 +0000 UTC m=+887.762059579" lastFinishedPulling="2026-03-09 18:40:17.257722769 +0000 UTC m=+898.600195167" observedRunningTime="2026-03-09 18:40:18.519116291 +0000 UTC m=+899.861588689" watchObservedRunningTime="2026-03-09 18:40:18.523013245 +0000 UTC m=+899.865485643" Mar 09 18:40:19 crc kubenswrapper[4750]: I0309 18:40:19.388768 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" path="/var/lib/kubelet/pods/29de4cb5-a1ae-446a-8e27-195aecca6b1a/volumes" Mar 09 18:40:20 crc kubenswrapper[4750]: I0309 18:40:20.323134 4750 scope.go:117] "RemoveContainer" containerID="0956a3e4f74a7e5ce542709aef043452900c4a885aa3b666b24560e27d248940" Mar 09 18:40:21 crc kubenswrapper[4750]: I0309 18:40:21.743560 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:40:21 crc kubenswrapper[4750]: I0309 18:40:21.744077 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:40:26 crc kubenswrapper[4750]: I0309 18:40:26.200444 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-7cmcp" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.890196 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6"] Mar 09 18:40:43 crc kubenswrapper[4750]: E0309 18:40:43.891411 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="extract-content" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.891432 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="extract-content" Mar 09 18:40:43 crc kubenswrapper[4750]: E0309 18:40:43.891463 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="registry-server" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.891475 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="registry-server" Mar 09 18:40:43 crc kubenswrapper[4750]: E0309 18:40:43.891490 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="extract-utilities" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.891501 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="extract-utilities" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.891695 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="29de4cb5-a1ae-446a-8e27-195aecca6b1a" containerName="registry-server" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.892969 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.895658 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.903530 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6"] Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.940236 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.940356 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:43 crc kubenswrapper[4750]: I0309 18:40:43.940401 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp5xc\" (UniqueName: \"kubernetes.io/projected/84a0fae0-78e5-4901-814b-e691c6b1805f-kube-api-access-vp5xc\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.041750 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp5xc\" (UniqueName: \"kubernetes.io/projected/84a0fae0-78e5-4901-814b-e691c6b1805f-kube-api-access-vp5xc\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.041859 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.041909 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.042502 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.042560 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.064086 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp5xc\" (UniqueName: \"kubernetes.io/projected/84a0fae0-78e5-4901-814b-e691c6b1805f-kube-api-access-vp5xc\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.216104 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.492934 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6"] Mar 09 18:40:44 crc kubenswrapper[4750]: I0309 18:40:44.694465 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" event={"ID":"84a0fae0-78e5-4901-814b-e691c6b1805f","Type":"ContainerStarted","Data":"0862c0ae65c76e02a8f437e64a418e5d38e657986b6698d8bc44c92d97bc6f62"} Mar 09 18:40:46 crc kubenswrapper[4750]: I0309 18:40:46.710261 4750 generic.go:334] "Generic (PLEG): container finished" podID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerID="a05e9aaeda03bdd20eaf632264492e600deead68cee750179823c131f08dd151" exitCode=0 Mar 09 18:40:46 crc kubenswrapper[4750]: I0309 18:40:46.710324 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" event={"ID":"84a0fae0-78e5-4901-814b-e691c6b1805f","Type":"ContainerDied","Data":"a05e9aaeda03bdd20eaf632264492e600deead68cee750179823c131f08dd151"} Mar 09 18:40:49 crc kubenswrapper[4750]: I0309 18:40:49.736605 4750 generic.go:334] "Generic (PLEG): container finished" podID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerID="d902762dc7844ded44872612d576deebb65e63aa2685548c12da32ba620e5991" exitCode=0 Mar 09 18:40:49 crc kubenswrapper[4750]: I0309 18:40:49.736780 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" event={"ID":"84a0fae0-78e5-4901-814b-e691c6b1805f","Type":"ContainerDied","Data":"d902762dc7844ded44872612d576deebb65e63aa2685548c12da32ba620e5991"} Mar 09 18:40:51 crc kubenswrapper[4750]: I0309 18:40:51.743974 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:40:51 crc kubenswrapper[4750]: I0309 18:40:51.744362 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:40:51 crc kubenswrapper[4750]: I0309 18:40:51.744422 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:40:51 crc kubenswrapper[4750]: I0309 18:40:51.746102 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d522a3b191af7d8f923266bd379a6ce4a390d5d0273a9db4e49eaf7ceef3304f"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:40:51 crc kubenswrapper[4750]: I0309 18:40:51.746233 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://d522a3b191af7d8f923266bd379a6ce4a390d5d0273a9db4e49eaf7ceef3304f" gracePeriod=600 Mar 09 18:40:51 crc kubenswrapper[4750]: I0309 18:40:51.754368 4750 generic.go:334] "Generic (PLEG): container finished" podID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerID="7d7259656235eb23fbdcb7e949d0670c7b4f91ce0d92c37be4061e93b88581c5" exitCode=0 Mar 09 18:40:51 crc kubenswrapper[4750]: I0309 18:40:51.754455 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" event={"ID":"84a0fae0-78e5-4901-814b-e691c6b1805f","Type":"ContainerDied","Data":"7d7259656235eb23fbdcb7e949d0670c7b4f91ce0d92c37be4061e93b88581c5"} Mar 09 18:40:52 crc kubenswrapper[4750]: I0309 18:40:52.763204 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="d522a3b191af7d8f923266bd379a6ce4a390d5d0273a9db4e49eaf7ceef3304f" exitCode=0 Mar 09 18:40:52 crc kubenswrapper[4750]: I0309 18:40:52.763287 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"d522a3b191af7d8f923266bd379a6ce4a390d5d0273a9db4e49eaf7ceef3304f"} Mar 09 18:40:52 crc kubenswrapper[4750]: I0309 18:40:52.763711 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"998a8326795d648e01a6e195fe1e013eaedd975e8692a77e8f1a9a6dca44107f"} Mar 09 18:40:52 crc kubenswrapper[4750]: I0309 18:40:52.763735 4750 scope.go:117] "RemoveContainer" containerID="dbc54b035b270a840e32e83b710affea17c5d6abaf9f561608508bb75a7047ba" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.043750 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.075538 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-bundle\") pod \"84a0fae0-78e5-4901-814b-e691c6b1805f\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.075737 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp5xc\" (UniqueName: \"kubernetes.io/projected/84a0fae0-78e5-4901-814b-e691c6b1805f-kube-api-access-vp5xc\") pod \"84a0fae0-78e5-4901-814b-e691c6b1805f\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.075793 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-util\") pod \"84a0fae0-78e5-4901-814b-e691c6b1805f\" (UID: \"84a0fae0-78e5-4901-814b-e691c6b1805f\") " Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.077849 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-bundle" (OuterVolumeSpecName: "bundle") pod "84a0fae0-78e5-4901-814b-e691c6b1805f" (UID: "84a0fae0-78e5-4901-814b-e691c6b1805f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.083844 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84a0fae0-78e5-4901-814b-e691c6b1805f-kube-api-access-vp5xc" (OuterVolumeSpecName: "kube-api-access-vp5xc") pod "84a0fae0-78e5-4901-814b-e691c6b1805f" (UID: "84a0fae0-78e5-4901-814b-e691c6b1805f"). InnerVolumeSpecName "kube-api-access-vp5xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.088706 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-util" (OuterVolumeSpecName: "util") pod "84a0fae0-78e5-4901-814b-e691c6b1805f" (UID: "84a0fae0-78e5-4901-814b-e691c6b1805f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.176492 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-util\") on node \"crc\" DevicePath \"\"" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.176567 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/84a0fae0-78e5-4901-814b-e691c6b1805f-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.176581 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp5xc\" (UniqueName: \"kubernetes.io/projected/84a0fae0-78e5-4901-814b-e691c6b1805f-kube-api-access-vp5xc\") on node \"crc\" DevicePath \"\"" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.774041 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" event={"ID":"84a0fae0-78e5-4901-814b-e691c6b1805f","Type":"ContainerDied","Data":"0862c0ae65c76e02a8f437e64a418e5d38e657986b6698d8bc44c92d97bc6f62"} Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.774117 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0862c0ae65c76e02a8f437e64a418e5d38e657986b6698d8bc44c92d97bc6f62" Mar 09 18:40:53 crc kubenswrapper[4750]: I0309 18:40:53.774170 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.438822 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj"] Mar 09 18:41:00 crc kubenswrapper[4750]: E0309 18:41:00.439863 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerName="pull" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.439881 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerName="pull" Mar 09 18:41:00 crc kubenswrapper[4750]: E0309 18:41:00.439898 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerName="extract" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.439906 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerName="extract" Mar 09 18:41:00 crc kubenswrapper[4750]: E0309 18:41:00.439921 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerName="util" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.439931 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerName="util" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.440074 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="84a0fae0-78e5-4901-814b-e691c6b1805f" containerName="extract" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.440702 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.442535 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.442552 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.444594 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4l4pl" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.455722 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj"] Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.487907 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqtfs\" (UniqueName: \"kubernetes.io/projected/a7132a9a-02c3-4361-9ccb-c135f416d289-kube-api-access-hqtfs\") pod \"nmstate-operator-75c5dccd6c-tk5qj\" (UID: \"a7132a9a-02c3-4361-9ccb-c135f416d289\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.588865 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqtfs\" (UniqueName: \"kubernetes.io/projected/a7132a9a-02c3-4361-9ccb-c135f416d289-kube-api-access-hqtfs\") pod \"nmstate-operator-75c5dccd6c-tk5qj\" (UID: \"a7132a9a-02c3-4361-9ccb-c135f416d289\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.621202 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqtfs\" (UniqueName: \"kubernetes.io/projected/a7132a9a-02c3-4361-9ccb-c135f416d289-kube-api-access-hqtfs\") pod \"nmstate-operator-75c5dccd6c-tk5qj\" (UID: \"a7132a9a-02c3-4361-9ccb-c135f416d289\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" Mar 09 18:41:00 crc kubenswrapper[4750]: I0309 18:41:00.797068 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" Mar 09 18:41:01 crc kubenswrapper[4750]: I0309 18:41:01.067916 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj"] Mar 09 18:41:01 crc kubenswrapper[4750]: I0309 18:41:01.080968 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 18:41:01 crc kubenswrapper[4750]: I0309 18:41:01.850161 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" event={"ID":"a7132a9a-02c3-4361-9ccb-c135f416d289","Type":"ContainerStarted","Data":"683aacab62f8012d85e38e167e2e79ddec1ef0b84ec0d592b9d2624f2e386716"} Mar 09 18:41:03 crc kubenswrapper[4750]: I0309 18:41:03.872174 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" event={"ID":"a7132a9a-02c3-4361-9ccb-c135f416d289","Type":"ContainerStarted","Data":"e91bd7c8d5edbfd985be6c7e7e5a7b3fd921dd54c757c1d92b653082441c3340"} Mar 09 18:41:03 crc kubenswrapper[4750]: I0309 18:41:03.902251 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-tk5qj" podStartSLOduration=1.442645247 podStartE2EDuration="3.902230665s" podCreationTimestamp="2026-03-09 18:41:00 +0000 UTC" firstStartedPulling="2026-03-09 18:41:01.080513867 +0000 UTC m=+942.422986275" lastFinishedPulling="2026-03-09 18:41:03.540099295 +0000 UTC m=+944.882571693" observedRunningTime="2026-03-09 18:41:03.893557881 +0000 UTC m=+945.236030309" watchObservedRunningTime="2026-03-09 18:41:03.902230665 +0000 UTC m=+945.244703073" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.153003 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-xzqjf"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.154664 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.160301 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-7tcq6" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.161021 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.162118 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.164089 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.192306 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-xzqjf"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.199314 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.216980 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-b5lff"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.218072 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.268047 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/73e03f8a-161c-4019-a06f-41e71db3f4d4-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-xtfm7\" (UID: \"73e03f8a-161c-4019-a06f-41e71db3f4d4\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.268129 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24n6r\" (UniqueName: \"kubernetes.io/projected/73e03f8a-161c-4019-a06f-41e71db3f4d4-kube-api-access-24n6r\") pod \"nmstate-webhook-786f45cff4-xtfm7\" (UID: \"73e03f8a-161c-4019-a06f-41e71db3f4d4\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.268246 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjlkg\" (UniqueName: \"kubernetes.io/projected/5eee205a-209f-4f39-8341-e5bfb1da1374-kube-api-access-gjlkg\") pod \"nmstate-metrics-69594cc75-xzqjf\" (UID: \"5eee205a-209f-4f39-8341-e5bfb1da1374\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.331772 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.332918 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.336279 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.336306 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.336792 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vnq9p" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.345074 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.369023 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjlkg\" (UniqueName: \"kubernetes.io/projected/5eee205a-209f-4f39-8341-e5bfb1da1374-kube-api-access-gjlkg\") pod \"nmstate-metrics-69594cc75-xzqjf\" (UID: \"5eee205a-209f-4f39-8341-e5bfb1da1374\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.369105 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbb4p\" (UniqueName: \"kubernetes.io/projected/b66128f4-2824-4e2a-93ec-4a59f313cb8d-kube-api-access-cbb4p\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.369139 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/73e03f8a-161c-4019-a06f-41e71db3f4d4-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-xtfm7\" (UID: \"73e03f8a-161c-4019-a06f-41e71db3f4d4\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.369161 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24n6r\" (UniqueName: \"kubernetes.io/projected/73e03f8a-161c-4019-a06f-41e71db3f4d4-kube-api-access-24n6r\") pod \"nmstate-webhook-786f45cff4-xtfm7\" (UID: \"73e03f8a-161c-4019-a06f-41e71db3f4d4\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.369182 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-dbus-socket\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.369198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-ovs-socket\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.369218 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-nmstate-lock\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.377137 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/73e03f8a-161c-4019-a06f-41e71db3f4d4-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-xtfm7\" (UID: \"73e03f8a-161c-4019-a06f-41e71db3f4d4\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.390981 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjlkg\" (UniqueName: \"kubernetes.io/projected/5eee205a-209f-4f39-8341-e5bfb1da1374-kube-api-access-gjlkg\") pod \"nmstate-metrics-69594cc75-xzqjf\" (UID: \"5eee205a-209f-4f39-8341-e5bfb1da1374\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.398742 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24n6r\" (UniqueName: \"kubernetes.io/projected/73e03f8a-161c-4019-a06f-41e71db3f4d4-kube-api-access-24n6r\") pod \"nmstate-webhook-786f45cff4-xtfm7\" (UID: \"73e03f8a-161c-4019-a06f-41e71db3f4d4\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471179 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-dbus-socket\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471253 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-ovs-socket\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471318 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-nmstate-lock\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471386 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471413 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-ovs-socket\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471447 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-nmstate-lock\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471418 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zckxl\" (UniqueName: \"kubernetes.io/projected/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-kube-api-access-zckxl\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471613 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b66128f4-2824-4e2a-93ec-4a59f313cb8d-dbus-socket\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471788 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbb4p\" (UniqueName: \"kubernetes.io/projected/b66128f4-2824-4e2a-93ec-4a59f313cb8d-kube-api-access-cbb4p\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.471817 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.489756 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbb4p\" (UniqueName: \"kubernetes.io/projected/b66128f4-2824-4e2a-93ec-4a59f313cb8d-kube-api-access-cbb4p\") pod \"nmstate-handler-b5lff\" (UID: \"b66128f4-2824-4e2a-93ec-4a59f313cb8d\") " pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.491213 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.510062 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.518812 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-788c78f896-9xcrb"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.519606 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.541057 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-788c78f896-9xcrb"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.549399 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.574661 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.574732 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.574754 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zckxl\" (UniqueName: \"kubernetes.io/projected/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-kube-api-access-zckxl\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.576342 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.592157 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.598843 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zckxl\" (UniqueName: \"kubernetes.io/projected/b3d9d77f-c39e-4e33-a33b-b75a6a9615ef-kube-api-access-zckxl\") pod \"nmstate-console-plugin-5dcbbd79cf-fg9ks\" (UID: \"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.653072 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.676415 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b768c8c1-312d-4a24-96fa-d76409332e5f-console-serving-cert\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.676470 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-trusted-ca-bundle\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.676504 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b768c8c1-312d-4a24-96fa-d76409332e5f-console-oauth-config\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.676523 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-service-ca\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.676545 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-oauth-serving-cert\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.676563 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpwrv\" (UniqueName: \"kubernetes.io/projected/b768c8c1-312d-4a24-96fa-d76409332e5f-kube-api-access-tpwrv\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.676597 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-console-config\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.765598 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7"] Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.778992 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-console-config\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.780491 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b768c8c1-312d-4a24-96fa-d76409332e5f-console-serving-cert\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.780339 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-console-config\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.780521 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-trusted-ca-bundle\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.780675 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b768c8c1-312d-4a24-96fa-d76409332e5f-console-oauth-config\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.780712 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-service-ca\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.780764 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-oauth-serving-cert\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.780797 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpwrv\" (UniqueName: \"kubernetes.io/projected/b768c8c1-312d-4a24-96fa-d76409332e5f-kube-api-access-tpwrv\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.781287 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-trusted-ca-bundle\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.782304 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-service-ca\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.784474 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b768c8c1-312d-4a24-96fa-d76409332e5f-oauth-serving-cert\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.788394 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b768c8c1-312d-4a24-96fa-d76409332e5f-console-serving-cert\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.788396 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b768c8c1-312d-4a24-96fa-d76409332e5f-console-oauth-config\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.796569 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpwrv\" (UniqueName: \"kubernetes.io/projected/b768c8c1-312d-4a24-96fa-d76409332e5f-kube-api-access-tpwrv\") pod \"console-788c78f896-9xcrb\" (UID: \"b768c8c1-312d-4a24-96fa-d76409332e5f\") " pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.891812 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b5lff" event={"ID":"b66128f4-2824-4e2a-93ec-4a59f313cb8d","Type":"ContainerStarted","Data":"ebe5c39bbc0420358dab1a04a4b10191f95af910ff1e0b9088015ac4e8ae1f18"} Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.893190 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" event={"ID":"73e03f8a-161c-4019-a06f-41e71db3f4d4","Type":"ContainerStarted","Data":"0cd6290ffe3ed0d5d1320c748430d8fe08d18cff90b192baf9e76722546aa033"} Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.895752 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:05 crc kubenswrapper[4750]: I0309 18:41:05.903541 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks"] Mar 09 18:41:06 crc kubenswrapper[4750]: I0309 18:41:06.048542 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-xzqjf"] Mar 09 18:41:06 crc kubenswrapper[4750]: I0309 18:41:06.205641 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-788c78f896-9xcrb"] Mar 09 18:41:06 crc kubenswrapper[4750]: W0309 18:41:06.213009 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb768c8c1_312d_4a24_96fa_d76409332e5f.slice/crio-4bd55b73d56a24fe70735677d812d6171721724e20c91031c501f32542f8b211 WatchSource:0}: Error finding container 4bd55b73d56a24fe70735677d812d6171721724e20c91031c501f32542f8b211: Status 404 returned error can't find the container with id 4bd55b73d56a24fe70735677d812d6171721724e20c91031c501f32542f8b211 Mar 09 18:41:06 crc kubenswrapper[4750]: I0309 18:41:06.904684 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" event={"ID":"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef","Type":"ContainerStarted","Data":"fd18ffe0b8b02d7e64b8e2860ae24fd32838efa2292c4c946b3818f8506079c9"} Mar 09 18:41:06 crc kubenswrapper[4750]: I0309 18:41:06.906886 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-788c78f896-9xcrb" event={"ID":"b768c8c1-312d-4a24-96fa-d76409332e5f","Type":"ContainerStarted","Data":"c6705e93ff9728e66e609302efab6f204d7fd33881ff30e31b56316e969494a9"} Mar 09 18:41:06 crc kubenswrapper[4750]: I0309 18:41:06.906930 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-788c78f896-9xcrb" event={"ID":"b768c8c1-312d-4a24-96fa-d76409332e5f","Type":"ContainerStarted","Data":"4bd55b73d56a24fe70735677d812d6171721724e20c91031c501f32542f8b211"} Mar 09 18:41:06 crc kubenswrapper[4750]: I0309 18:41:06.910786 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" event={"ID":"5eee205a-209f-4f39-8341-e5bfb1da1374","Type":"ContainerStarted","Data":"117dcb367274eeb22074fecf1142f861af9f46e32c87c02bb91d0730d909c113"} Mar 09 18:41:06 crc kubenswrapper[4750]: I0309 18:41:06.938437 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-788c78f896-9xcrb" podStartSLOduration=1.938408462 podStartE2EDuration="1.938408462s" podCreationTimestamp="2026-03-09 18:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:41:06.931311821 +0000 UTC m=+948.273784229" watchObservedRunningTime="2026-03-09 18:41:06.938408462 +0000 UTC m=+948.280880870" Mar 09 18:41:09 crc kubenswrapper[4750]: I0309 18:41:09.937146 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b5lff" event={"ID":"b66128f4-2824-4e2a-93ec-4a59f313cb8d","Type":"ContainerStarted","Data":"0ce4dc98bbbe4f799214556a0598699dfb06cafe3834b641ccffcfb485f5a69f"} Mar 09 18:41:09 crc kubenswrapper[4750]: I0309 18:41:09.938119 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:09 crc kubenswrapper[4750]: I0309 18:41:09.940800 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" event={"ID":"5eee205a-209f-4f39-8341-e5bfb1da1374","Type":"ContainerStarted","Data":"27f81eb97896311310cbeb49ebd0e918292d562a72ebdb53e568d960dc393e1e"} Mar 09 18:41:09 crc kubenswrapper[4750]: I0309 18:41:09.944795 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" event={"ID":"73e03f8a-161c-4019-a06f-41e71db3f4d4","Type":"ContainerStarted","Data":"3210dfd17e70ebba56492a52e8b7e89e0e6efe3c81753b17f105934adcb33c69"} Mar 09 18:41:09 crc kubenswrapper[4750]: I0309 18:41:09.944992 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:09 crc kubenswrapper[4750]: I0309 18:41:09.958131 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-b5lff" podStartSLOduration=1.842396371 podStartE2EDuration="4.958118306s" podCreationTimestamp="2026-03-09 18:41:05 +0000 UTC" firstStartedPulling="2026-03-09 18:41:05.585175291 +0000 UTC m=+946.927647679" lastFinishedPulling="2026-03-09 18:41:08.700897176 +0000 UTC m=+950.043369614" observedRunningTime="2026-03-09 18:41:09.953858251 +0000 UTC m=+951.296330689" watchObservedRunningTime="2026-03-09 18:41:09.958118306 +0000 UTC m=+951.300590704" Mar 09 18:41:09 crc kubenswrapper[4750]: I0309 18:41:09.975302 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" podStartSLOduration=2.1079411 podStartE2EDuration="4.975287669s" podCreationTimestamp="2026-03-09 18:41:05 +0000 UTC" firstStartedPulling="2026-03-09 18:41:05.787446472 +0000 UTC m=+947.129918860" lastFinishedPulling="2026-03-09 18:41:08.654793031 +0000 UTC m=+949.997265429" observedRunningTime="2026-03-09 18:41:09.974581801 +0000 UTC m=+951.317054219" watchObservedRunningTime="2026-03-09 18:41:09.975287669 +0000 UTC m=+951.317760077" Mar 09 18:41:10 crc kubenswrapper[4750]: I0309 18:41:10.958598 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" event={"ID":"b3d9d77f-c39e-4e33-a33b-b75a6a9615ef","Type":"ContainerStarted","Data":"c2f5927452ebe58398f943179a4642383e33b03c9b15ff8f9eb216e5faee5d69"} Mar 09 18:41:10 crc kubenswrapper[4750]: I0309 18:41:10.984863 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-fg9ks" podStartSLOduration=1.727267002 podStartE2EDuration="5.984837971s" podCreationTimestamp="2026-03-09 18:41:05 +0000 UTC" firstStartedPulling="2026-03-09 18:41:05.91250724 +0000 UTC m=+947.254979638" lastFinishedPulling="2026-03-09 18:41:10.170078209 +0000 UTC m=+951.512550607" observedRunningTime="2026-03-09 18:41:10.981682896 +0000 UTC m=+952.324155304" watchObservedRunningTime="2026-03-09 18:41:10.984837971 +0000 UTC m=+952.327310369" Mar 09 18:41:11 crc kubenswrapper[4750]: I0309 18:41:11.966405 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" event={"ID":"5eee205a-209f-4f39-8341-e5bfb1da1374","Type":"ContainerStarted","Data":"30898d7c2bba11741d09c47e62c124b64f5c62e64009a73f870468e0087f008d"} Mar 09 18:41:11 crc kubenswrapper[4750]: I0309 18:41:11.989729 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-69594cc75-xzqjf" podStartSLOduration=1.357106616 podStartE2EDuration="6.989707137s" podCreationTimestamp="2026-03-09 18:41:05 +0000 UTC" firstStartedPulling="2026-03-09 18:41:06.069790647 +0000 UTC m=+947.412263045" lastFinishedPulling="2026-03-09 18:41:11.702391168 +0000 UTC m=+953.044863566" observedRunningTime="2026-03-09 18:41:11.985621826 +0000 UTC m=+953.328094224" watchObservedRunningTime="2026-03-09 18:41:11.989707137 +0000 UTC m=+953.332179535" Mar 09 18:41:15 crc kubenswrapper[4750]: I0309 18:41:15.582940 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-b5lff" Mar 09 18:41:15 crc kubenswrapper[4750]: I0309 18:41:15.896443 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:15 crc kubenswrapper[4750]: I0309 18:41:15.896517 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:15 crc kubenswrapper[4750]: I0309 18:41:15.902937 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:16 crc kubenswrapper[4750]: I0309 18:41:16.002443 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-788c78f896-9xcrb" Mar 09 18:41:16 crc kubenswrapper[4750]: I0309 18:41:16.064806 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bglg2"] Mar 09 18:41:25 crc kubenswrapper[4750]: I0309 18:41:25.519084 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-786f45cff4-xtfm7" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.125874 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-bglg2" podUID="863600af-6300-46d6-875a-85b0622c8f50" containerName="console" containerID="cri-o://71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1" gracePeriod=15 Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.582987 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bglg2_863600af-6300-46d6-875a-85b0622c8f50/console/0.log" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.583070 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.684679 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-oauth-config\") pod \"863600af-6300-46d6-875a-85b0622c8f50\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.684793 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-service-ca\") pod \"863600af-6300-46d6-875a-85b0622c8f50\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.684893 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-console-config\") pod \"863600af-6300-46d6-875a-85b0622c8f50\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.684960 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-serving-cert\") pod \"863600af-6300-46d6-875a-85b0622c8f50\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.684988 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-trusted-ca-bundle\") pod \"863600af-6300-46d6-875a-85b0622c8f50\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.685040 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-oauth-serving-cert\") pod \"863600af-6300-46d6-875a-85b0622c8f50\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.685070 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhxmn\" (UniqueName: \"kubernetes.io/projected/863600af-6300-46d6-875a-85b0622c8f50-kube-api-access-jhxmn\") pod \"863600af-6300-46d6-875a-85b0622c8f50\" (UID: \"863600af-6300-46d6-875a-85b0622c8f50\") " Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.686955 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-service-ca" (OuterVolumeSpecName: "service-ca") pod "863600af-6300-46d6-875a-85b0622c8f50" (UID: "863600af-6300-46d6-875a-85b0622c8f50"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.687038 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "863600af-6300-46d6-875a-85b0622c8f50" (UID: "863600af-6300-46d6-875a-85b0622c8f50"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.687108 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-console-config" (OuterVolumeSpecName: "console-config") pod "863600af-6300-46d6-875a-85b0622c8f50" (UID: "863600af-6300-46d6-875a-85b0622c8f50"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.687271 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "863600af-6300-46d6-875a-85b0622c8f50" (UID: "863600af-6300-46d6-875a-85b0622c8f50"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.693680 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "863600af-6300-46d6-875a-85b0622c8f50" (UID: "863600af-6300-46d6-875a-85b0622c8f50"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.695004 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/863600af-6300-46d6-875a-85b0622c8f50-kube-api-access-jhxmn" (OuterVolumeSpecName: "kube-api-access-jhxmn") pod "863600af-6300-46d6-875a-85b0622c8f50" (UID: "863600af-6300-46d6-875a-85b0622c8f50"). InnerVolumeSpecName "kube-api-access-jhxmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.695657 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "863600af-6300-46d6-875a-85b0622c8f50" (UID: "863600af-6300-46d6-875a-85b0622c8f50"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.786224 4750 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.786278 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.786290 4750 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.786301 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhxmn\" (UniqueName: \"kubernetes.io/projected/863600af-6300-46d6-875a-85b0622c8f50-kube-api-access-jhxmn\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.786313 4750 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/863600af-6300-46d6-875a-85b0622c8f50-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.786321 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:41 crc kubenswrapper[4750]: I0309 18:41:41.786330 4750 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/863600af-6300-46d6-875a-85b0622c8f50-console-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.221543 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bglg2_863600af-6300-46d6-875a-85b0622c8f50/console/0.log" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.221688 4750 generic.go:334] "Generic (PLEG): container finished" podID="863600af-6300-46d6-875a-85b0622c8f50" containerID="71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1" exitCode=2 Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.221795 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bglg2" event={"ID":"863600af-6300-46d6-875a-85b0622c8f50","Type":"ContainerDied","Data":"71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1"} Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.221830 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bglg2" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.221855 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bglg2" event={"ID":"863600af-6300-46d6-875a-85b0622c8f50","Type":"ContainerDied","Data":"8a0387233f74db99d6e02b6cbfa634ae69172f4c272ca750011d1b0cf121ca47"} Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.221896 4750 scope.go:117] "RemoveContainer" containerID="71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.256939 4750 scope.go:117] "RemoveContainer" containerID="71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1" Mar 09 18:41:42 crc kubenswrapper[4750]: E0309 18:41:42.258336 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1\": container with ID starting with 71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1 not found: ID does not exist" containerID="71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.258403 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1"} err="failed to get container status \"71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1\": rpc error: code = NotFound desc = could not find container \"71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1\": container with ID starting with 71fa9c9f16ba63b01426df15f454deb5b1dd27fe862487efd6eeaec8d4a7e4d1 not found: ID does not exist" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.264488 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bglg2"] Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.271868 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-bglg2"] Mar 09 18:41:42 crc kubenswrapper[4750]: E0309 18:41:42.280342 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod863600af_6300_46d6_875a_85b0622c8f50.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod863600af_6300_46d6_875a_85b0622c8f50.slice/crio-8a0387233f74db99d6e02b6cbfa634ae69172f4c272ca750011d1b0cf121ca47\": RecentStats: unable to find data in memory cache]" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.747049 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc"] Mar 09 18:41:42 crc kubenswrapper[4750]: E0309 18:41:42.747659 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863600af-6300-46d6-875a-85b0622c8f50" containerName="console" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.747674 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="863600af-6300-46d6-875a-85b0622c8f50" containerName="console" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.747801 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="863600af-6300-46d6-875a-85b0622c8f50" containerName="console" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.748726 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.751431 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.773999 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc"] Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.903159 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.903245 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwt7s\" (UniqueName: \"kubernetes.io/projected/bf52966f-af17-425e-a64b-109c9245bcec-kube-api-access-gwt7s\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:42 crc kubenswrapper[4750]: I0309 18:41:42.903284 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.004188 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt7s\" (UniqueName: \"kubernetes.io/projected/bf52966f-af17-425e-a64b-109c9245bcec-kube-api-access-gwt7s\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.004316 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.004945 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.005086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.005455 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.026778 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwt7s\" (UniqueName: \"kubernetes.io/projected/bf52966f-af17-425e-a64b-109c9245bcec-kube-api-access-gwt7s\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.065137 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.385940 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="863600af-6300-46d6-875a-85b0622c8f50" path="/var/lib/kubelet/pods/863600af-6300-46d6-875a-85b0622c8f50/volumes" Mar 09 18:41:43 crc kubenswrapper[4750]: I0309 18:41:43.428597 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc"] Mar 09 18:41:44 crc kubenswrapper[4750]: I0309 18:41:44.478165 4750 generic.go:334] "Generic (PLEG): container finished" podID="bf52966f-af17-425e-a64b-109c9245bcec" containerID="db26b46eb7552201185dc6b3fcf9cb06b8a56f220e788b9ae487c4d787f24a03" exitCode=0 Mar 09 18:41:44 crc kubenswrapper[4750]: I0309 18:41:44.478493 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" event={"ID":"bf52966f-af17-425e-a64b-109c9245bcec","Type":"ContainerDied","Data":"db26b46eb7552201185dc6b3fcf9cb06b8a56f220e788b9ae487c4d787f24a03"} Mar 09 18:41:44 crc kubenswrapper[4750]: I0309 18:41:44.478536 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" event={"ID":"bf52966f-af17-425e-a64b-109c9245bcec","Type":"ContainerStarted","Data":"89785f6613fbe523ac2bb0ef7abfe3c5719a8b61653c400d73bceaae08b94268"} Mar 09 18:41:50 crc kubenswrapper[4750]: I0309 18:41:50.524531 4750 generic.go:334] "Generic (PLEG): container finished" podID="bf52966f-af17-425e-a64b-109c9245bcec" containerID="ffbe91f8999d31f955ec33d2a5502c040c7a2f0d1ba584769acd5acf57d66e2f" exitCode=0 Mar 09 18:41:50 crc kubenswrapper[4750]: I0309 18:41:50.524699 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" event={"ID":"bf52966f-af17-425e-a64b-109c9245bcec","Type":"ContainerDied","Data":"ffbe91f8999d31f955ec33d2a5502c040c7a2f0d1ba584769acd5acf57d66e2f"} Mar 09 18:41:51 crc kubenswrapper[4750]: I0309 18:41:51.533498 4750 generic.go:334] "Generic (PLEG): container finished" podID="bf52966f-af17-425e-a64b-109c9245bcec" containerID="09c8387ad9d98a5e6138c2b9917c91aa6f96751d9b95e3d4aa88a81aaa7d489d" exitCode=0 Mar 09 18:41:51 crc kubenswrapper[4750]: I0309 18:41:51.533563 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" event={"ID":"bf52966f-af17-425e-a64b-109c9245bcec","Type":"ContainerDied","Data":"09c8387ad9d98a5e6138c2b9917c91aa6f96751d9b95e3d4aa88a81aaa7d489d"} Mar 09 18:41:52 crc kubenswrapper[4750]: I0309 18:41:52.846698 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.040768 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-util\") pod \"bf52966f-af17-425e-a64b-109c9245bcec\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.040875 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-bundle\") pod \"bf52966f-af17-425e-a64b-109c9245bcec\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.040940 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwt7s\" (UniqueName: \"kubernetes.io/projected/bf52966f-af17-425e-a64b-109c9245bcec-kube-api-access-gwt7s\") pod \"bf52966f-af17-425e-a64b-109c9245bcec\" (UID: \"bf52966f-af17-425e-a64b-109c9245bcec\") " Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.042082 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-bundle" (OuterVolumeSpecName: "bundle") pod "bf52966f-af17-425e-a64b-109c9245bcec" (UID: "bf52966f-af17-425e-a64b-109c9245bcec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.050970 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-util" (OuterVolumeSpecName: "util") pod "bf52966f-af17-425e-a64b-109c9245bcec" (UID: "bf52966f-af17-425e-a64b-109c9245bcec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.051612 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf52966f-af17-425e-a64b-109c9245bcec-kube-api-access-gwt7s" (OuterVolumeSpecName: "kube-api-access-gwt7s") pod "bf52966f-af17-425e-a64b-109c9245bcec" (UID: "bf52966f-af17-425e-a64b-109c9245bcec"). InnerVolumeSpecName "kube-api-access-gwt7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.142691 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.142760 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwt7s\" (UniqueName: \"kubernetes.io/projected/bf52966f-af17-425e-a64b-109c9245bcec-kube-api-access-gwt7s\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.142787 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf52966f-af17-425e-a64b-109c9245bcec-util\") on node \"crc\" DevicePath \"\"" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.555741 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" event={"ID":"bf52966f-af17-425e-a64b-109c9245bcec","Type":"ContainerDied","Data":"89785f6613fbe523ac2bb0ef7abfe3c5719a8b61653c400d73bceaae08b94268"} Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.555825 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89785f6613fbe523ac2bb0ef7abfe3c5719a8b61653c400d73bceaae08b94268" Mar 09 18:41:53 crc kubenswrapper[4750]: I0309 18:41:53.555949 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.132976 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551362-kgfhj"] Mar 09 18:42:00 crc kubenswrapper[4750]: E0309 18:42:00.134150 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf52966f-af17-425e-a64b-109c9245bcec" containerName="util" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.134180 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf52966f-af17-425e-a64b-109c9245bcec" containerName="util" Mar 09 18:42:00 crc kubenswrapper[4750]: E0309 18:42:00.134225 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf52966f-af17-425e-a64b-109c9245bcec" containerName="pull" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.134243 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf52966f-af17-425e-a64b-109c9245bcec" containerName="pull" Mar 09 18:42:00 crc kubenswrapper[4750]: E0309 18:42:00.134262 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf52966f-af17-425e-a64b-109c9245bcec" containerName="extract" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.134279 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf52966f-af17-425e-a64b-109c9245bcec" containerName="extract" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.134527 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf52966f-af17-425e-a64b-109c9245bcec" containerName="extract" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.135555 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551362-kgfhj" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.144529 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551362-kgfhj"] Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.146015 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.146015 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.149413 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.159806 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4cnv\" (UniqueName: \"kubernetes.io/projected/b622dcef-e250-4105-bb3c-245620f16c39-kube-api-access-z4cnv\") pod \"auto-csr-approver-29551362-kgfhj\" (UID: \"b622dcef-e250-4105-bb3c-245620f16c39\") " pod="openshift-infra/auto-csr-approver-29551362-kgfhj" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.264229 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4cnv\" (UniqueName: \"kubernetes.io/projected/b622dcef-e250-4105-bb3c-245620f16c39-kube-api-access-z4cnv\") pod \"auto-csr-approver-29551362-kgfhj\" (UID: \"b622dcef-e250-4105-bb3c-245620f16c39\") " pod="openshift-infra/auto-csr-approver-29551362-kgfhj" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.294817 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4cnv\" (UniqueName: \"kubernetes.io/projected/b622dcef-e250-4105-bb3c-245620f16c39-kube-api-access-z4cnv\") pod \"auto-csr-approver-29551362-kgfhj\" (UID: \"b622dcef-e250-4105-bb3c-245620f16c39\") " pod="openshift-infra/auto-csr-approver-29551362-kgfhj" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.461240 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551362-kgfhj" Mar 09 18:42:00 crc kubenswrapper[4750]: I0309 18:42:00.702123 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551362-kgfhj"] Mar 09 18:42:01 crc kubenswrapper[4750]: I0309 18:42:01.656340 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551362-kgfhj" event={"ID":"b622dcef-e250-4105-bb3c-245620f16c39","Type":"ContainerStarted","Data":"6d453cfd20f530efeb79fb86a1168f1a43543c62885cfb1e8c06412831238d72"} Mar 09 18:42:02 crc kubenswrapper[4750]: I0309 18:42:02.666250 4750 generic.go:334] "Generic (PLEG): container finished" podID="b622dcef-e250-4105-bb3c-245620f16c39" containerID="475b8bc49d7bc443a4a1a95a0b3ede890a53eea2c72957ed06a64305689d8abc" exitCode=0 Mar 09 18:42:02 crc kubenswrapper[4750]: I0309 18:42:02.666348 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551362-kgfhj" event={"ID":"b622dcef-e250-4105-bb3c-245620f16c39","Type":"ContainerDied","Data":"475b8bc49d7bc443a4a1a95a0b3ede890a53eea2c72957ed06a64305689d8abc"} Mar 09 18:42:04 crc kubenswrapper[4750]: I0309 18:42:04.030551 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551362-kgfhj" Mar 09 18:42:04 crc kubenswrapper[4750]: I0309 18:42:04.221371 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4cnv\" (UniqueName: \"kubernetes.io/projected/b622dcef-e250-4105-bb3c-245620f16c39-kube-api-access-z4cnv\") pod \"b622dcef-e250-4105-bb3c-245620f16c39\" (UID: \"b622dcef-e250-4105-bb3c-245620f16c39\") " Mar 09 18:42:04 crc kubenswrapper[4750]: I0309 18:42:04.227522 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b622dcef-e250-4105-bb3c-245620f16c39-kube-api-access-z4cnv" (OuterVolumeSpecName: "kube-api-access-z4cnv") pod "b622dcef-e250-4105-bb3c-245620f16c39" (UID: "b622dcef-e250-4105-bb3c-245620f16c39"). InnerVolumeSpecName "kube-api-access-z4cnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:42:04 crc kubenswrapper[4750]: I0309 18:42:04.324070 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4cnv\" (UniqueName: \"kubernetes.io/projected/b622dcef-e250-4105-bb3c-245620f16c39-kube-api-access-z4cnv\") on node \"crc\" DevicePath \"\"" Mar 09 18:42:04 crc kubenswrapper[4750]: I0309 18:42:04.699504 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551362-kgfhj" event={"ID":"b622dcef-e250-4105-bb3c-245620f16c39","Type":"ContainerDied","Data":"6d453cfd20f530efeb79fb86a1168f1a43543c62885cfb1e8c06412831238d72"} Mar 09 18:42:04 crc kubenswrapper[4750]: I0309 18:42:04.699555 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d453cfd20f530efeb79fb86a1168f1a43543c62885cfb1e8c06412831238d72" Mar 09 18:42:04 crc kubenswrapper[4750]: I0309 18:42:04.699596 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551362-kgfhj" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.133182 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551356-w5mg4"] Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.137439 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551356-w5mg4"] Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.384373 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="957dd1c4-3eac-4206-b18e-a337e54ca850" path="/var/lib/kubelet/pods/957dd1c4-3eac-4206-b18e-a337e54ca850/volumes" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.935829 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp"] Mar 09 18:42:05 crc kubenswrapper[4750]: E0309 18:42:05.936422 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b622dcef-e250-4105-bb3c-245620f16c39" containerName="oc" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.936447 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b622dcef-e250-4105-bb3c-245620f16c39" containerName="oc" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.936557 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b622dcef-e250-4105-bb3c-245620f16c39" containerName="oc" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.937052 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:05 crc kubenswrapper[4750]: W0309 18:42:05.938817 4750 reflector.go:561] object-"metallb-system"/"metallb-operator-controller-manager-service-cert": failed to list *v1.Secret: secrets "metallb-operator-controller-manager-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Mar 09 18:42:05 crc kubenswrapper[4750]: E0309 18:42:05.938906 4750 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-controller-manager-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-controller-manager-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:42:05 crc kubenswrapper[4750]: W0309 18:42:05.939208 4750 reflector.go:561] object-"metallb-system"/"metallb-operator-webhook-server-cert": failed to list *v1.Secret: secrets "metallb-operator-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Mar 09 18:42:05 crc kubenswrapper[4750]: E0309 18:42:05.939232 4750 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.939689 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7ltnx" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.940515 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.940705 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 09 18:42:05 crc kubenswrapper[4750]: I0309 18:42:05.954791 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp"] Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.047276 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-webhook-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.047695 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnbtd\" (UniqueName: \"kubernetes.io/projected/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-kube-api-access-cnbtd\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.047888 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-apiservice-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.148432 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-webhook-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.148505 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnbtd\" (UniqueName: \"kubernetes.io/projected/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-kube-api-access-cnbtd\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.148543 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-apiservice-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.168606 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnbtd\" (UniqueName: \"kubernetes.io/projected/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-kube-api-access-cnbtd\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.289053 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz"] Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.290048 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.292782 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.292834 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-87j7x" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.293005 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.304909 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz"] Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.453102 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1687d828-b0e8-45ef-8828-00af0b7ea22f-apiservice-cert\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.453229 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1687d828-b0e8-45ef-8828-00af0b7ea22f-webhook-cert\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.453261 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltls8\" (UniqueName: \"kubernetes.io/projected/1687d828-b0e8-45ef-8828-00af0b7ea22f-kube-api-access-ltls8\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.554407 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1687d828-b0e8-45ef-8828-00af0b7ea22f-webhook-cert\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.554482 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltls8\" (UniqueName: \"kubernetes.io/projected/1687d828-b0e8-45ef-8828-00af0b7ea22f-kube-api-access-ltls8\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.554526 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1687d828-b0e8-45ef-8828-00af0b7ea22f-apiservice-cert\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.560001 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1687d828-b0e8-45ef-8828-00af0b7ea22f-webhook-cert\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.562224 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1687d828-b0e8-45ef-8828-00af0b7ea22f-apiservice-cert\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.580220 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltls8\" (UniqueName: \"kubernetes.io/projected/1687d828-b0e8-45ef-8828-00af0b7ea22f-kube-api-access-ltls8\") pod \"metallb-operator-webhook-server-749bcdd774-cjqbz\" (UID: \"1687d828-b0e8-45ef-8828-00af0b7ea22f\") " pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.605028 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:06 crc kubenswrapper[4750]: I0309 18:42:06.942914 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz"] Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.084153 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 09 18:42:07 crc kubenswrapper[4750]: E0309 18:42:07.149069 4750 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 18:42:07 crc kubenswrapper[4750]: E0309 18:42:07.149114 4750 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 18:42:07 crc kubenswrapper[4750]: E0309 18:42:07.149193 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-apiservice-cert podName:7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd nodeName:}" failed. No retries permitted until 2026-03-09 18:42:07.649167341 +0000 UTC m=+1008.991639739 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-apiservice-cert") pod "metallb-operator-controller-manager-6754f6499c-wbkgp" (UID: "7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:42:07 crc kubenswrapper[4750]: E0309 18:42:07.149212 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-webhook-cert podName:7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd nodeName:}" failed. No retries permitted until 2026-03-09 18:42:07.649205743 +0000 UTC m=+1008.991678141 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-webhook-cert") pod "metallb-operator-controller-manager-6754f6499c-wbkgp" (UID: "7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd") : failed to sync secret cache: timed out waiting for the condition Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.169230 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.673873 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-webhook-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.674081 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-apiservice-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.681257 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-webhook-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.684497 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd-apiservice-cert\") pod \"metallb-operator-controller-manager-6754f6499c-wbkgp\" (UID: \"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd\") " pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.751987 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:07 crc kubenswrapper[4750]: I0309 18:42:07.754380 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" event={"ID":"1687d828-b0e8-45ef-8828-00af0b7ea22f","Type":"ContainerStarted","Data":"11920a25aef557818cd6f58f2d6e69998c031d0922fa0fcd69fc225d3e2d3649"} Mar 09 18:42:08 crc kubenswrapper[4750]: I0309 18:42:08.035247 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp"] Mar 09 18:42:08 crc kubenswrapper[4750]: W0309 18:42:08.035881 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7878a0a9_d982_4ea1_bc04_e2c2bee5e7fd.slice/crio-a2966103ad003472c852564aaaf5e39174890cfa65afad058f668ba37908f931 WatchSource:0}: Error finding container a2966103ad003472c852564aaaf5e39174890cfa65afad058f668ba37908f931: Status 404 returned error can't find the container with id a2966103ad003472c852564aaaf5e39174890cfa65afad058f668ba37908f931 Mar 09 18:42:08 crc kubenswrapper[4750]: I0309 18:42:08.765294 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" event={"ID":"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd","Type":"ContainerStarted","Data":"a2966103ad003472c852564aaaf5e39174890cfa65afad058f668ba37908f931"} Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.141288 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qzqws"] Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.142906 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.155080 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzqws"] Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.245318 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-utilities\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.245416 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-catalog-content\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.245482 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmjcw\" (UniqueName: \"kubernetes.io/projected/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-kube-api-access-mmjcw\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.346855 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmjcw\" (UniqueName: \"kubernetes.io/projected/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-kube-api-access-mmjcw\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.347019 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-utilities\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.347104 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-catalog-content\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.347705 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-utilities\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.347992 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-catalog-content\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.388863 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmjcw\" (UniqueName: \"kubernetes.io/projected/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-kube-api-access-mmjcw\") pod \"redhat-marketplace-qzqws\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:12 crc kubenswrapper[4750]: I0309 18:42:12.470563 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.472119 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzqws"] Mar 09 18:42:13 crc kubenswrapper[4750]: W0309 18:42:13.489019 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fab58f0_2c16_4048_bbb2_e2a9d7d480dd.slice/crio-96137d2a68f91292cbf158276449c5183ec3de441f01d65ea30d0fb58838351f WatchSource:0}: Error finding container 96137d2a68f91292cbf158276449c5183ec3de441f01d65ea30d0fb58838351f: Status 404 returned error can't find the container with id 96137d2a68f91292cbf158276449c5183ec3de441f01d65ea30d0fb58838351f Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.838447 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" event={"ID":"1687d828-b0e8-45ef-8828-00af0b7ea22f","Type":"ContainerStarted","Data":"d6e357641f88cffd70b2fc87d8220ac8227288b7ecbc2755e4c2f54201dbeaf4"} Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.838973 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.840956 4750 generic.go:334] "Generic (PLEG): container finished" podID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerID="acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1" exitCode=0 Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.841014 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzqws" event={"ID":"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd","Type":"ContainerDied","Data":"acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1"} Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.841033 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzqws" event={"ID":"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd","Type":"ContainerStarted","Data":"96137d2a68f91292cbf158276449c5183ec3de441f01d65ea30d0fb58838351f"} Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.843554 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" event={"ID":"7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd","Type":"ContainerStarted","Data":"a574ca17c3b00a0b9043bbbf9031459b66675d63fc46434819587ca1a6084a85"} Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.843804 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.861603 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" podStartSLOduration=1.708274509 podStartE2EDuration="7.861584432s" podCreationTimestamp="2026-03-09 18:42:06 +0000 UTC" firstStartedPulling="2026-03-09 18:42:06.96286981 +0000 UTC m=+1008.305342208" lastFinishedPulling="2026-03-09 18:42:13.116179733 +0000 UTC m=+1014.458652131" observedRunningTime="2026-03-09 18:42:13.858324064 +0000 UTC m=+1015.200796462" watchObservedRunningTime="2026-03-09 18:42:13.861584432 +0000 UTC m=+1015.204056830" Mar 09 18:42:13 crc kubenswrapper[4750]: I0309 18:42:13.899014 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" podStartSLOduration=3.799731183 podStartE2EDuration="8.898990652s" podCreationTimestamp="2026-03-09 18:42:05 +0000 UTC" firstStartedPulling="2026-03-09 18:42:08.038816466 +0000 UTC m=+1009.381288864" lastFinishedPulling="2026-03-09 18:42:13.138075925 +0000 UTC m=+1014.480548333" observedRunningTime="2026-03-09 18:42:13.898668133 +0000 UTC m=+1015.241140521" watchObservedRunningTime="2026-03-09 18:42:13.898990652 +0000 UTC m=+1015.241463050" Mar 09 18:42:14 crc kubenswrapper[4750]: I0309 18:42:14.853411 4750 generic.go:334] "Generic (PLEG): container finished" podID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerID="369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a" exitCode=0 Mar 09 18:42:14 crc kubenswrapper[4750]: I0309 18:42:14.853479 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzqws" event={"ID":"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd","Type":"ContainerDied","Data":"369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a"} Mar 09 18:42:15 crc kubenswrapper[4750]: I0309 18:42:15.863135 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzqws" event={"ID":"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd","Type":"ContainerStarted","Data":"bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010"} Mar 09 18:42:15 crc kubenswrapper[4750]: I0309 18:42:15.887204 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qzqws" podStartSLOduration=2.3622828829999998 podStartE2EDuration="3.887177761s" podCreationTimestamp="2026-03-09 18:42:12 +0000 UTC" firstStartedPulling="2026-03-09 18:42:13.842199659 +0000 UTC m=+1015.184672057" lastFinishedPulling="2026-03-09 18:42:15.367094537 +0000 UTC m=+1016.709566935" observedRunningTime="2026-03-09 18:42:15.880565653 +0000 UTC m=+1017.223038071" watchObservedRunningTime="2026-03-09 18:42:15.887177761 +0000 UTC m=+1017.229650149" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.315267 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jqgfx"] Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.316660 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.326665 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-catalog-content\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.326794 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2tq6\" (UniqueName: \"kubernetes.io/projected/9941b051-c194-49fd-afc9-1b7748329619-kube-api-access-s2tq6\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.326826 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-utilities\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.331430 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jqgfx"] Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.427806 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2tq6\" (UniqueName: \"kubernetes.io/projected/9941b051-c194-49fd-afc9-1b7748329619-kube-api-access-s2tq6\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.427861 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-utilities\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.427911 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-catalog-content\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.428437 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-catalog-content\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.428624 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-utilities\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.449900 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2tq6\" (UniqueName: \"kubernetes.io/projected/9941b051-c194-49fd-afc9-1b7748329619-kube-api-access-s2tq6\") pod \"community-operators-jqgfx\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:16 crc kubenswrapper[4750]: I0309 18:42:16.636298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:17 crc kubenswrapper[4750]: I0309 18:42:17.005747 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jqgfx"] Mar 09 18:42:17 crc kubenswrapper[4750]: I0309 18:42:17.891019 4750 generic.go:334] "Generic (PLEG): container finished" podID="9941b051-c194-49fd-afc9-1b7748329619" containerID="1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e" exitCode=0 Mar 09 18:42:17 crc kubenswrapper[4750]: I0309 18:42:17.891142 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqgfx" event={"ID":"9941b051-c194-49fd-afc9-1b7748329619","Type":"ContainerDied","Data":"1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e"} Mar 09 18:42:17 crc kubenswrapper[4750]: I0309 18:42:17.891413 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqgfx" event={"ID":"9941b051-c194-49fd-afc9-1b7748329619","Type":"ContainerStarted","Data":"6ab8002ae5e78a4ac12e1a79264acb739bf954da67b37107ea422e32cb4d44e5"} Mar 09 18:42:18 crc kubenswrapper[4750]: I0309 18:42:18.903565 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqgfx" event={"ID":"9941b051-c194-49fd-afc9-1b7748329619","Type":"ContainerStarted","Data":"dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51"} Mar 09 18:42:19 crc kubenswrapper[4750]: I0309 18:42:19.912239 4750 generic.go:334] "Generic (PLEG): container finished" podID="9941b051-c194-49fd-afc9-1b7748329619" containerID="dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51" exitCode=0 Mar 09 18:42:19 crc kubenswrapper[4750]: I0309 18:42:19.912392 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqgfx" event={"ID":"9941b051-c194-49fd-afc9-1b7748329619","Type":"ContainerDied","Data":"dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51"} Mar 09 18:42:20 crc kubenswrapper[4750]: I0309 18:42:20.568426 4750 scope.go:117] "RemoveContainer" containerID="e6f7f0768d424f3fce16f78de9d11dfd4696c27337a61d96238e62647e39a3ae" Mar 09 18:42:20 crc kubenswrapper[4750]: I0309 18:42:20.921898 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqgfx" event={"ID":"9941b051-c194-49fd-afc9-1b7748329619","Type":"ContainerStarted","Data":"58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad"} Mar 09 18:42:20 crc kubenswrapper[4750]: I0309 18:42:20.940849 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jqgfx" podStartSLOduration=2.155847174 podStartE2EDuration="4.940828499s" podCreationTimestamp="2026-03-09 18:42:16 +0000 UTC" firstStartedPulling="2026-03-09 18:42:17.895071741 +0000 UTC m=+1019.237544139" lastFinishedPulling="2026-03-09 18:42:20.680053066 +0000 UTC m=+1022.022525464" observedRunningTime="2026-03-09 18:42:20.938930817 +0000 UTC m=+1022.281403225" watchObservedRunningTime="2026-03-09 18:42:20.940828499 +0000 UTC m=+1022.283300887" Mar 09 18:42:22 crc kubenswrapper[4750]: I0309 18:42:22.472797 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:22 crc kubenswrapper[4750]: I0309 18:42:22.473179 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:22 crc kubenswrapper[4750]: I0309 18:42:22.528951 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:22 crc kubenswrapper[4750]: I0309 18:42:22.995755 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.107793 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzqws"] Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.108374 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qzqws" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="registry-server" containerID="cri-o://bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010" gracePeriod=2 Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.588942 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.617945 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-749bcdd774-cjqbz" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.637313 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.637369 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.692217 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.696369 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-utilities\") pod \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.696455 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmjcw\" (UniqueName: \"kubernetes.io/projected/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-kube-api-access-mmjcw\") pod \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.696523 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-catalog-content\") pod \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\" (UID: \"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd\") " Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.698175 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-utilities" (OuterVolumeSpecName: "utilities") pod "4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" (UID: "4fab58f0-2c16-4048-bbb2-e2a9d7d480dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.726932 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-kube-api-access-mmjcw" (OuterVolumeSpecName: "kube-api-access-mmjcw") pod "4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" (UID: "4fab58f0-2c16-4048-bbb2-e2a9d7d480dd"). InnerVolumeSpecName "kube-api-access-mmjcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.738111 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" (UID: "4fab58f0-2c16-4048-bbb2-e2a9d7d480dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.798727 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.798784 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmjcw\" (UniqueName: \"kubernetes.io/projected/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-kube-api-access-mmjcw\") on node \"crc\" DevicePath \"\"" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.798797 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.963229 4750 generic.go:334] "Generic (PLEG): container finished" podID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerID="bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010" exitCode=0 Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.963295 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzqws" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.963320 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzqws" event={"ID":"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd","Type":"ContainerDied","Data":"bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010"} Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.963777 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzqws" event={"ID":"4fab58f0-2c16-4048-bbb2-e2a9d7d480dd","Type":"ContainerDied","Data":"96137d2a68f91292cbf158276449c5183ec3de441f01d65ea30d0fb58838351f"} Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.963832 4750 scope.go:117] "RemoveContainer" containerID="bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010" Mar 09 18:42:26 crc kubenswrapper[4750]: I0309 18:42:26.996960 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzqws"] Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.000021 4750 scope.go:117] "RemoveContainer" containerID="369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.003178 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzqws"] Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.036015 4750 scope.go:117] "RemoveContainer" containerID="acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.043100 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.077208 4750 scope.go:117] "RemoveContainer" containerID="bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010" Mar 09 18:42:27 crc kubenswrapper[4750]: E0309 18:42:27.077793 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010\": container with ID starting with bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010 not found: ID does not exist" containerID="bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.077858 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010"} err="failed to get container status \"bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010\": rpc error: code = NotFound desc = could not find container \"bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010\": container with ID starting with bcd8672ca134a136c57b538594933b8120b0020611da8f8eb9fcd9ff944dd010 not found: ID does not exist" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.077898 4750 scope.go:117] "RemoveContainer" containerID="369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a" Mar 09 18:42:27 crc kubenswrapper[4750]: E0309 18:42:27.078290 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a\": container with ID starting with 369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a not found: ID does not exist" containerID="369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.078415 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a"} err="failed to get container status \"369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a\": rpc error: code = NotFound desc = could not find container \"369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a\": container with ID starting with 369cc9144ac176e44331f157cef7c903a1041878876f0d5201ead8e17b60ac2a not found: ID does not exist" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.078520 4750 scope.go:117] "RemoveContainer" containerID="acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1" Mar 09 18:42:27 crc kubenswrapper[4750]: E0309 18:42:27.078949 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1\": container with ID starting with acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1 not found: ID does not exist" containerID="acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.078990 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1"} err="failed to get container status \"acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1\": rpc error: code = NotFound desc = could not find container \"acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1\": container with ID starting with acc59bb9aabe78c52cefd2782251d501c7e0058093b89b9e9288bdd01c6705e1 not found: ID does not exist" Mar 09 18:42:27 crc kubenswrapper[4750]: I0309 18:42:27.386887 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" path="/var/lib/kubelet/pods/4fab58f0-2c16-4048-bbb2-e2a9d7d480dd/volumes" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.305923 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jqgfx"] Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.306493 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jqgfx" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="registry-server" containerID="cri-o://58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad" gracePeriod=2 Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.677208 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.754896 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-utilities\") pod \"9941b051-c194-49fd-afc9-1b7748329619\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.754987 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-catalog-content\") pod \"9941b051-c194-49fd-afc9-1b7748329619\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.755012 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2tq6\" (UniqueName: \"kubernetes.io/projected/9941b051-c194-49fd-afc9-1b7748329619-kube-api-access-s2tq6\") pod \"9941b051-c194-49fd-afc9-1b7748329619\" (UID: \"9941b051-c194-49fd-afc9-1b7748329619\") " Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.756972 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-utilities" (OuterVolumeSpecName: "utilities") pod "9941b051-c194-49fd-afc9-1b7748329619" (UID: "9941b051-c194-49fd-afc9-1b7748329619"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.764174 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9941b051-c194-49fd-afc9-1b7748329619-kube-api-access-s2tq6" (OuterVolumeSpecName: "kube-api-access-s2tq6") pod "9941b051-c194-49fd-afc9-1b7748329619" (UID: "9941b051-c194-49fd-afc9-1b7748329619"). InnerVolumeSpecName "kube-api-access-s2tq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.836027 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9941b051-c194-49fd-afc9-1b7748329619" (UID: "9941b051-c194-49fd-afc9-1b7748329619"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.856293 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.856820 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9941b051-c194-49fd-afc9-1b7748329619-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.856914 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2tq6\" (UniqueName: \"kubernetes.io/projected/9941b051-c194-49fd-afc9-1b7748329619-kube-api-access-s2tq6\") on node \"crc\" DevicePath \"\"" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.996231 4750 generic.go:334] "Generic (PLEG): container finished" podID="9941b051-c194-49fd-afc9-1b7748329619" containerID="58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad" exitCode=0 Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.996293 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqgfx" event={"ID":"9941b051-c194-49fd-afc9-1b7748329619","Type":"ContainerDied","Data":"58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad"} Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.996335 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqgfx" event={"ID":"9941b051-c194-49fd-afc9-1b7748329619","Type":"ContainerDied","Data":"6ab8002ae5e78a4ac12e1a79264acb739bf954da67b37107ea422e32cb4d44e5"} Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.996358 4750 scope.go:117] "RemoveContainer" containerID="58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad" Mar 09 18:42:30 crc kubenswrapper[4750]: I0309 18:42:30.996799 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqgfx" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.016258 4750 scope.go:117] "RemoveContainer" containerID="dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.030269 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jqgfx"] Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.037096 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jqgfx"] Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.053019 4750 scope.go:117] "RemoveContainer" containerID="1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.073166 4750 scope.go:117] "RemoveContainer" containerID="58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad" Mar 09 18:42:31 crc kubenswrapper[4750]: E0309 18:42:31.074259 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad\": container with ID starting with 58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad not found: ID does not exist" containerID="58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.074387 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad"} err="failed to get container status \"58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad\": rpc error: code = NotFound desc = could not find container \"58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad\": container with ID starting with 58de5612233118113f7e4208cabd1be38ece076ec2107e608354e78ef4a7caad not found: ID does not exist" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.074470 4750 scope.go:117] "RemoveContainer" containerID="dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51" Mar 09 18:42:31 crc kubenswrapper[4750]: E0309 18:42:31.075192 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51\": container with ID starting with dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51 not found: ID does not exist" containerID="dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.075296 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51"} err="failed to get container status \"dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51\": rpc error: code = NotFound desc = could not find container \"dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51\": container with ID starting with dcdcb9337894d5d7b97cde458f16639e46a7e57cc5bd217790bafaca203e8c51 not found: ID does not exist" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.075366 4750 scope.go:117] "RemoveContainer" containerID="1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e" Mar 09 18:42:31 crc kubenswrapper[4750]: E0309 18:42:31.077837 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e\": container with ID starting with 1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e not found: ID does not exist" containerID="1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.077865 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e"} err="failed to get container status \"1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e\": rpc error: code = NotFound desc = could not find container \"1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e\": container with ID starting with 1f192bb28db55342ffd802ae71db939e33391fd14e93ddff1420263c2731732e not found: ID does not exist" Mar 09 18:42:31 crc kubenswrapper[4750]: I0309 18:42:31.409407 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9941b051-c194-49fd-afc9-1b7748329619" path="/var/lib/kubelet/pods/9941b051-c194-49fd-afc9-1b7748329619/volumes" Mar 09 18:42:47 crc kubenswrapper[4750]: I0309 18:42:47.754813 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6754f6499c-wbkgp" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.620175 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-spjm2"] Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.620887 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="registry-server" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.620908 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="registry-server" Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.620918 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="extract-content" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.620925 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="extract-content" Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.620933 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="extract-utilities" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.620940 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="extract-utilities" Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.620950 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="registry-server" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.620956 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="registry-server" Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.620963 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="extract-utilities" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.620969 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="extract-utilities" Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.620994 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="extract-content" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.621003 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="extract-content" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.621134 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9941b051-c194-49fd-afc9-1b7748329619" containerName="registry-server" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.621147 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fab58f0-2c16-4048-bbb2-e2a9d7d480dd" containerName="registry-server" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.623412 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.625691 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj"] Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.626644 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.627868 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.627934 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.628084 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xfs8c" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.630538 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.644448 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj"] Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.735404 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bqt4d"] Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.736707 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.739422 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.739598 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.739734 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.739853 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-tb992" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.758413 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-86ddb6bd46-d9bmt"] Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759203 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-frr-sockets\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759263 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d818f1c-4170-4157-b621-fb422f11c109-metrics-certs\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759292 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-metrics\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759329 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-reloader\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759369 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-frr-conf\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759397 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4rzt\" (UniqueName: \"kubernetes.io/projected/9d818f1c-4170-4157-b621-fb422f11c109-kube-api-access-g4rzt\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759416 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dk9l\" (UniqueName: \"kubernetes.io/projected/ef9fe80b-913b-424a-a780-8223aed6eb50-kube-api-access-4dk9l\") pod \"frr-k8s-webhook-server-7f989f654f-5s5xj\" (UID: \"ef9fe80b-913b-424a-a780-8223aed6eb50\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759443 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ef9fe80b-913b-424a-a780-8223aed6eb50-cert\") pod \"frr-k8s-webhook-server-7f989f654f-5s5xj\" (UID: \"ef9fe80b-913b-424a-a780-8223aed6eb50\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.759459 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9d818f1c-4170-4157-b621-fb422f11c109-frr-startup\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.760247 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.766513 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.771678 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-d9bmt"] Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.860383 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-metrics\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.860427 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-reloader\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.860457 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.860498 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-frr-conf\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.860529 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4rzt\" (UniqueName: \"kubernetes.io/projected/9d818f1c-4170-4157-b621-fb422f11c109-kube-api-access-g4rzt\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.861417 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-frr-conf\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.861475 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-metrics\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.861583 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dk9l\" (UniqueName: \"kubernetes.io/projected/ef9fe80b-913b-424a-a780-8223aed6eb50-kube-api-access-4dk9l\") pod \"frr-k8s-webhook-server-7f989f654f-5s5xj\" (UID: \"ef9fe80b-913b-424a-a780-8223aed6eb50\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.861787 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-reloader\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.862065 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhmzv\" (UniqueName: \"kubernetes.io/projected/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-kube-api-access-dhmzv\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.862140 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ef9fe80b-913b-424a-a780-8223aed6eb50-cert\") pod \"frr-k8s-webhook-server-7f989f654f-5s5xj\" (UID: \"ef9fe80b-913b-424a-a780-8223aed6eb50\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.862176 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9d818f1c-4170-4157-b621-fb422f11c109-frr-startup\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.862264 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-metrics-certs\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.863166 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9d818f1c-4170-4157-b621-fb422f11c109-frr-startup\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.863226 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-metallb-excludel2\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.863250 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-frr-sockets\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.863379 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d818f1c-4170-4157-b621-fb422f11c109-metrics-certs\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.863733 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9d818f1c-4170-4157-b621-fb422f11c109-frr-sockets\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.869486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d818f1c-4170-4157-b621-fb422f11c109-metrics-certs\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.880315 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dk9l\" (UniqueName: \"kubernetes.io/projected/ef9fe80b-913b-424a-a780-8223aed6eb50-kube-api-access-4dk9l\") pod \"frr-k8s-webhook-server-7f989f654f-5s5xj\" (UID: \"ef9fe80b-913b-424a-a780-8223aed6eb50\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.883234 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ef9fe80b-913b-424a-a780-8223aed6eb50-cert\") pod \"frr-k8s-webhook-server-7f989f654f-5s5xj\" (UID: \"ef9fe80b-913b-424a-a780-8223aed6eb50\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.905721 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4rzt\" (UniqueName: \"kubernetes.io/projected/9d818f1c-4170-4157-b621-fb422f11c109-kube-api-access-g4rzt\") pod \"frr-k8s-spjm2\" (UID: \"9d818f1c-4170-4157-b621-fb422f11c109\") " pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.965115 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-metrics-certs\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.965180 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-metallb-excludel2\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.965209 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39f2ab0a-439c-404f-872e-e40e746282d1-cert\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.965274 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.965296 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/39f2ab0a-439c-404f-872e-e40e746282d1-metrics-certs\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.965352 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhmzv\" (UniqueName: \"kubernetes.io/projected/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-kube-api-access-dhmzv\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.965398 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv48v\" (UniqueName: \"kubernetes.io/projected/39f2ab0a-439c-404f-872e-e40e746282d1-kube-api-access-rv48v\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.965522 4750 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 09 18:42:48 crc kubenswrapper[4750]: E0309 18:42:48.965607 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist podName:4d0613aa-6eb9-43d8-bf29-e63e6b053c7f nodeName:}" failed. No retries permitted until 2026-03-09 18:42:49.465564678 +0000 UTC m=+1050.808037086 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist") pod "speaker-bqt4d" (UID: "4d0613aa-6eb9-43d8-bf29-e63e6b053c7f") : secret "metallb-memberlist" not found Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.966063 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-metallb-excludel2\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.972087 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-metrics-certs\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.974165 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-spjm2" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.981474 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:48 crc kubenswrapper[4750]: I0309 18:42:48.985256 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhmzv\" (UniqueName: \"kubernetes.io/projected/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-kube-api-access-dhmzv\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.066841 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv48v\" (UniqueName: \"kubernetes.io/projected/39f2ab0a-439c-404f-872e-e40e746282d1-kube-api-access-rv48v\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.067280 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39f2ab0a-439c-404f-872e-e40e746282d1-cert\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.067356 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/39f2ab0a-439c-404f-872e-e40e746282d1-metrics-certs\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.071447 4750 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.073951 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/39f2ab0a-439c-404f-872e-e40e746282d1-metrics-certs\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.088879 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39f2ab0a-439c-404f-872e-e40e746282d1-cert\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.091371 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv48v\" (UniqueName: \"kubernetes.io/projected/39f2ab0a-439c-404f-872e-e40e746282d1-kube-api-access-rv48v\") pod \"controller-86ddb6bd46-d9bmt\" (UID: \"39f2ab0a-439c-404f-872e-e40e746282d1\") " pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.190380 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerStarted","Data":"7c5d63dc430930884789a590c03c93f7b421add6c618051bee3e47142827b8aa"} Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.266181 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj"] Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.382395 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.473223 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:49 crc kubenswrapper[4750]: E0309 18:42:49.473430 4750 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 09 18:42:49 crc kubenswrapper[4750]: E0309 18:42:49.473502 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist podName:4d0613aa-6eb9-43d8-bf29-e63e6b053c7f nodeName:}" failed. No retries permitted until 2026-03-09 18:42:50.473479814 +0000 UTC m=+1051.815952212 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist") pod "speaker-bqt4d" (UID: "4d0613aa-6eb9-43d8-bf29-e63e6b053c7f") : secret "metallb-memberlist" not found Mar 09 18:42:49 crc kubenswrapper[4750]: I0309 18:42:49.613939 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-d9bmt"] Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.198446 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" event={"ID":"ef9fe80b-913b-424a-a780-8223aed6eb50","Type":"ContainerStarted","Data":"a839c49e755d95f90ad834cff6dff8940774fd01b6a3e2919f212c61e759f6a2"} Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.200756 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-d9bmt" event={"ID":"39f2ab0a-439c-404f-872e-e40e746282d1","Type":"ContainerStarted","Data":"1dd05e07ee4e4ebc70134682a5681e1ec2af9531a71da90660ed7cca3e13859b"} Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.200813 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-d9bmt" event={"ID":"39f2ab0a-439c-404f-872e-e40e746282d1","Type":"ContainerStarted","Data":"122b57d3b994eb26da37aea2a14e686be92d8e967d1ead3044fe4aa174ed20eb"} Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.200825 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-d9bmt" event={"ID":"39f2ab0a-439c-404f-872e-e40e746282d1","Type":"ContainerStarted","Data":"f2bf46f8de20c32d4a535dc97c3405812abb9e25f7a90655d64e41597c6fd876"} Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.201067 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.229604 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-86ddb6bd46-d9bmt" podStartSLOduration=2.229575543 podStartE2EDuration="2.229575543s" podCreationTimestamp="2026-03-09 18:42:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:42:50.229354117 +0000 UTC m=+1051.571826555" watchObservedRunningTime="2026-03-09 18:42:50.229575543 +0000 UTC m=+1051.572047951" Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.488970 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.500784 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4d0613aa-6eb9-43d8-bf29-e63e6b053c7f-memberlist\") pod \"speaker-bqt4d\" (UID: \"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f\") " pod="metallb-system/speaker-bqt4d" Mar 09 18:42:50 crc kubenswrapper[4750]: I0309 18:42:50.551475 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bqt4d" Mar 09 18:42:51 crc kubenswrapper[4750]: I0309 18:42:51.210221 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bqt4d" event={"ID":"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f","Type":"ContainerStarted","Data":"51d65bc13e0080518a9a3af6e52d2057059f9c4c39ce8d0fb52643c5a953e006"} Mar 09 18:42:51 crc kubenswrapper[4750]: I0309 18:42:51.210286 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bqt4d" event={"ID":"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f","Type":"ContainerStarted","Data":"bbd10ee1f100a9a537c57dd5277824533d639dadf550e144d86098448ee0334c"} Mar 09 18:42:51 crc kubenswrapper[4750]: I0309 18:42:51.743948 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:42:51 crc kubenswrapper[4750]: I0309 18:42:51.744028 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.228141 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bqt4d" event={"ID":"4d0613aa-6eb9-43d8-bf29-e63e6b053c7f","Type":"ContainerStarted","Data":"3c8b2f7fac10132f745719f344a9c7cc70df51e9e3a1177757366a0746dd8fa7"} Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.229191 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bqt4d" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.331129 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bqt4d" podStartSLOduration=4.331101369 podStartE2EDuration="4.331101369s" podCreationTimestamp="2026-03-09 18:42:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:42:52.263343753 +0000 UTC m=+1053.605816151" watchObservedRunningTime="2026-03-09 18:42:52.331101369 +0000 UTC m=+1053.673573767" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.331996 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t6fdh"] Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.333262 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.349172 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t6fdh"] Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.520198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-994gl\" (UniqueName: \"kubernetes.io/projected/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-kube-api-access-994gl\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.520371 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-utilities\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.520397 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-catalog-content\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.621499 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-utilities\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.621549 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-catalog-content\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.621591 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-994gl\" (UniqueName: \"kubernetes.io/projected/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-kube-api-access-994gl\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.622100 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-catalog-content\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.622471 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-utilities\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.662735 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-994gl\" (UniqueName: \"kubernetes.io/projected/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-kube-api-access-994gl\") pod \"certified-operators-t6fdh\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:52 crc kubenswrapper[4750]: I0309 18:42:52.948208 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:42:53 crc kubenswrapper[4750]: I0309 18:42:53.434043 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t6fdh"] Mar 09 18:42:54 crc kubenswrapper[4750]: I0309 18:42:54.246258 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerID="5f04f21caadb790c47c5c75276e75741958b35675d3d285693ad5bc09fd259af" exitCode=0 Mar 09 18:42:54 crc kubenswrapper[4750]: I0309 18:42:54.246378 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6fdh" event={"ID":"e2242148-3cf5-4dd7-97f3-ed10f483a2b2","Type":"ContainerDied","Data":"5f04f21caadb790c47c5c75276e75741958b35675d3d285693ad5bc09fd259af"} Mar 09 18:42:54 crc kubenswrapper[4750]: I0309 18:42:54.246413 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6fdh" event={"ID":"e2242148-3cf5-4dd7-97f3-ed10f483a2b2","Type":"ContainerStarted","Data":"df111622a8ff23db77f63dc092608177b775863836966edfa9217c9ed07b3f35"} Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.389205 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-86ddb6bd46-d9bmt" Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.713539 4750 generic.go:334] "Generic (PLEG): container finished" podID="9d818f1c-4170-4157-b621-fb422f11c109" containerID="d5f899c803f50fa62050cfda4ae2b623b5f8abd56767ed6ccb8868aa8bd72f84" exitCode=0 Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.713680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerDied","Data":"d5f899c803f50fa62050cfda4ae2b623b5f8abd56767ed6ccb8868aa8bd72f84"} Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.717470 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerID="4320354f4424b6c104225c1be7d144c0432544e0477776a21c781a1b1836621d" exitCode=0 Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.717613 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6fdh" event={"ID":"e2242148-3cf5-4dd7-97f3-ed10f483a2b2","Type":"ContainerDied","Data":"4320354f4424b6c104225c1be7d144c0432544e0477776a21c781a1b1836621d"} Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.719837 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" event={"ID":"ef9fe80b-913b-424a-a780-8223aed6eb50","Type":"ContainerStarted","Data":"0f63630d0aab09f990452f997ec14585bae12a970c1a21ac26eb6359b6b35ada"} Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.719988 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:42:59 crc kubenswrapper[4750]: I0309 18:42:59.776228 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" podStartSLOduration=2.353238102 podStartE2EDuration="11.776199599s" podCreationTimestamp="2026-03-09 18:42:48 +0000 UTC" firstStartedPulling="2026-03-09 18:42:49.279213826 +0000 UTC m=+1050.621686234" lastFinishedPulling="2026-03-09 18:42:58.702175333 +0000 UTC m=+1060.044647731" observedRunningTime="2026-03-09 18:42:59.770614129 +0000 UTC m=+1061.113086527" watchObservedRunningTime="2026-03-09 18:42:59.776199599 +0000 UTC m=+1061.118671997" Mar 09 18:43:00 crc kubenswrapper[4750]: I0309 18:43:00.557885 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bqt4d" Mar 09 18:43:00 crc kubenswrapper[4750]: I0309 18:43:00.730854 4750 generic.go:334] "Generic (PLEG): container finished" podID="9d818f1c-4170-4157-b621-fb422f11c109" containerID="ac22c4f40d18719ff9f128e80d5128ff9b0a41c98733b7790c753f4a9162baac" exitCode=0 Mar 09 18:43:00 crc kubenswrapper[4750]: I0309 18:43:00.730909 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerDied","Data":"ac22c4f40d18719ff9f128e80d5128ff9b0a41c98733b7790c753f4a9162baac"} Mar 09 18:43:00 crc kubenswrapper[4750]: I0309 18:43:00.734248 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6fdh" event={"ID":"e2242148-3cf5-4dd7-97f3-ed10f483a2b2","Type":"ContainerStarted","Data":"e89acd048bf50aa7970c0550c479a1a0440bf666115ca81d16ff2e803eef92f1"} Mar 09 18:43:00 crc kubenswrapper[4750]: I0309 18:43:00.794606 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t6fdh" podStartSLOduration=2.661993839 podStartE2EDuration="8.794580853s" podCreationTimestamp="2026-03-09 18:42:52 +0000 UTC" firstStartedPulling="2026-03-09 18:42:54.248139019 +0000 UTC m=+1055.590611417" lastFinishedPulling="2026-03-09 18:43:00.380726033 +0000 UTC m=+1061.723198431" observedRunningTime="2026-03-09 18:43:00.788841219 +0000 UTC m=+1062.131313627" watchObservedRunningTime="2026-03-09 18:43:00.794580853 +0000 UTC m=+1062.137053251" Mar 09 18:43:01 crc kubenswrapper[4750]: I0309 18:43:01.745962 4750 generic.go:334] "Generic (PLEG): container finished" podID="9d818f1c-4170-4157-b621-fb422f11c109" containerID="7a1fd8d223205bbc6ea2157374b3f1bd10ddd36c9537bafabe3b7de0b2f735bf" exitCode=0 Mar 09 18:43:01 crc kubenswrapper[4750]: I0309 18:43:01.746078 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerDied","Data":"7a1fd8d223205bbc6ea2157374b3f1bd10ddd36c9537bafabe3b7de0b2f735bf"} Mar 09 18:43:02 crc kubenswrapper[4750]: I0309 18:43:02.761316 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerStarted","Data":"2065f69016d6b7430b5b5078155b31027598d30e21e03da097e0f1fb94775ab5"} Mar 09 18:43:02 crc kubenswrapper[4750]: I0309 18:43:02.761715 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerStarted","Data":"2f28368ced729bf84411d6d5caba2fde17621474667a88c2a7cb9904d3fe6ff6"} Mar 09 18:43:02 crc kubenswrapper[4750]: I0309 18:43:02.761728 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerStarted","Data":"346e7f66799a7052daf4171e9297f596681b6da28c9b5cf2699f6008087b5c97"} Mar 09 18:43:02 crc kubenswrapper[4750]: I0309 18:43:02.761737 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerStarted","Data":"427725a7bd6810456d0f09784e177e0e85db4f709b3e6a4bf7e1ef968d488ee9"} Mar 09 18:43:02 crc kubenswrapper[4750]: I0309 18:43:02.761745 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerStarted","Data":"b34563d7bd208574c37b2cfbf3c472a319af5a181f9dd0842e74d48a2becd523"} Mar 09 18:43:02 crc kubenswrapper[4750]: I0309 18:43:02.949549 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:43:02 crc kubenswrapper[4750]: I0309 18:43:02.949609 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:43:03 crc kubenswrapper[4750]: I0309 18:43:03.001075 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:43:03 crc kubenswrapper[4750]: I0309 18:43:03.774166 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-spjm2" event={"ID":"9d818f1c-4170-4157-b621-fb422f11c109","Type":"ContainerStarted","Data":"a24b483dc7b8269ee4a3608effaca584b028ad08473b0c51a9b496ab4ead67d0"} Mar 09 18:43:03 crc kubenswrapper[4750]: I0309 18:43:03.774251 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-spjm2" Mar 09 18:43:03 crc kubenswrapper[4750]: I0309 18:43:03.807466 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-spjm2" podStartSLOduration=6.265150193 podStartE2EDuration="15.807440939s" podCreationTimestamp="2026-03-09 18:42:48 +0000 UTC" firstStartedPulling="2026-03-09 18:42:49.159576689 +0000 UTC m=+1050.502049077" lastFinishedPulling="2026-03-09 18:42:58.701867425 +0000 UTC m=+1060.044339823" observedRunningTime="2026-03-09 18:43:03.803273397 +0000 UTC m=+1065.145745785" watchObservedRunningTime="2026-03-09 18:43:03.807440939 +0000 UTC m=+1065.149913337" Mar 09 18:43:03 crc kubenswrapper[4750]: I0309 18:43:03.975495 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-spjm2" Mar 09 18:43:04 crc kubenswrapper[4750]: I0309 18:43:04.011961 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-spjm2" Mar 09 18:43:07 crc kubenswrapper[4750]: I0309 18:43:07.925073 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-q9qqf"] Mar 09 18:43:07 crc kubenswrapper[4750]: I0309 18:43:07.926729 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:07 crc kubenswrapper[4750]: I0309 18:43:07.929041 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-jtbnt" Mar 09 18:43:07 crc kubenswrapper[4750]: I0309 18:43:07.929585 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 09 18:43:07 crc kubenswrapper[4750]: I0309 18:43:07.932238 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 09 18:43:07 crc kubenswrapper[4750]: I0309 18:43:07.937152 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q9qqf"] Mar 09 18:43:08 crc kubenswrapper[4750]: I0309 18:43:08.083301 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pct87\" (UniqueName: \"kubernetes.io/projected/a08a8a31-630e-45ad-a319-c2904fa895b3-kube-api-access-pct87\") pod \"openstack-operator-index-q9qqf\" (UID: \"a08a8a31-630e-45ad-a319-c2904fa895b3\") " pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:08 crc kubenswrapper[4750]: I0309 18:43:08.185519 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pct87\" (UniqueName: \"kubernetes.io/projected/a08a8a31-630e-45ad-a319-c2904fa895b3-kube-api-access-pct87\") pod \"openstack-operator-index-q9qqf\" (UID: \"a08a8a31-630e-45ad-a319-c2904fa895b3\") " pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:08 crc kubenswrapper[4750]: I0309 18:43:08.221479 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pct87\" (UniqueName: \"kubernetes.io/projected/a08a8a31-630e-45ad-a319-c2904fa895b3-kube-api-access-pct87\") pod \"openstack-operator-index-q9qqf\" (UID: \"a08a8a31-630e-45ad-a319-c2904fa895b3\") " pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:08 crc kubenswrapper[4750]: I0309 18:43:08.249553 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:08 crc kubenswrapper[4750]: I0309 18:43:08.551846 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q9qqf"] Mar 09 18:43:08 crc kubenswrapper[4750]: W0309 18:43:08.560611 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda08a8a31_630e_45ad_a319_c2904fa895b3.slice/crio-4fa10e53a89d31944f5751970c7d1c00f66150ac7d4765d41572af2b1d38cba1 WatchSource:0}: Error finding container 4fa10e53a89d31944f5751970c7d1c00f66150ac7d4765d41572af2b1d38cba1: Status 404 returned error can't find the container with id 4fa10e53a89d31944f5751970c7d1c00f66150ac7d4765d41572af2b1d38cba1 Mar 09 18:43:08 crc kubenswrapper[4750]: I0309 18:43:08.816818 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q9qqf" event={"ID":"a08a8a31-630e-45ad-a319-c2904fa895b3","Type":"ContainerStarted","Data":"4fa10e53a89d31944f5751970c7d1c00f66150ac7d4765d41572af2b1d38cba1"} Mar 09 18:43:08 crc kubenswrapper[4750]: I0309 18:43:08.987580 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-5s5xj" Mar 09 18:43:12 crc kubenswrapper[4750]: I0309 18:43:12.854996 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q9qqf" event={"ID":"a08a8a31-630e-45ad-a319-c2904fa895b3","Type":"ContainerStarted","Data":"672c89ee459fffa59bd5a48a5ec384ddaa0cd1a673e6439d9f448bf00eb628be"} Mar 09 18:43:12 crc kubenswrapper[4750]: I0309 18:43:12.881582 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-q9qqf" podStartSLOduration=2.426342463 podStartE2EDuration="5.881520797s" podCreationTimestamp="2026-03-09 18:43:07 +0000 UTC" firstStartedPulling="2026-03-09 18:43:08.564555102 +0000 UTC m=+1069.907027510" lastFinishedPulling="2026-03-09 18:43:12.019733446 +0000 UTC m=+1073.362205844" observedRunningTime="2026-03-09 18:43:12.878940948 +0000 UTC m=+1074.221413386" watchObservedRunningTime="2026-03-09 18:43:12.881520797 +0000 UTC m=+1074.223993235" Mar 09 18:43:13 crc kubenswrapper[4750]: I0309 18:43:13.045112 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:43:17 crc kubenswrapper[4750]: I0309 18:43:17.714030 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t6fdh"] Mar 09 18:43:17 crc kubenswrapper[4750]: I0309 18:43:17.715869 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t6fdh" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="registry-server" containerID="cri-o://e89acd048bf50aa7970c0550c479a1a0440bf666115ca81d16ff2e803eef92f1" gracePeriod=2 Mar 09 18:43:17 crc kubenswrapper[4750]: I0309 18:43:17.898723 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerID="e89acd048bf50aa7970c0550c479a1a0440bf666115ca81d16ff2e803eef92f1" exitCode=0 Mar 09 18:43:17 crc kubenswrapper[4750]: I0309 18:43:17.898793 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6fdh" event={"ID":"e2242148-3cf5-4dd7-97f3-ed10f483a2b2","Type":"ContainerDied","Data":"e89acd048bf50aa7970c0550c479a1a0440bf666115ca81d16ff2e803eef92f1"} Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.128919 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.249114 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-catalog-content\") pod \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.249240 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-utilities\") pod \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.249349 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-994gl\" (UniqueName: \"kubernetes.io/projected/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-kube-api-access-994gl\") pod \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\" (UID: \"e2242148-3cf5-4dd7-97f3-ed10f483a2b2\") " Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.250453 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-utilities" (OuterVolumeSpecName: "utilities") pod "e2242148-3cf5-4dd7-97f3-ed10f483a2b2" (UID: "e2242148-3cf5-4dd7-97f3-ed10f483a2b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.250710 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.250755 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.264398 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-kube-api-access-994gl" (OuterVolumeSpecName: "kube-api-access-994gl") pod "e2242148-3cf5-4dd7-97f3-ed10f483a2b2" (UID: "e2242148-3cf5-4dd7-97f3-ed10f483a2b2"). InnerVolumeSpecName "kube-api-access-994gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.302993 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2242148-3cf5-4dd7-97f3-ed10f483a2b2" (UID: "e2242148-3cf5-4dd7-97f3-ed10f483a2b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.305870 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.350826 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-994gl\" (UniqueName: \"kubernetes.io/projected/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-kube-api-access-994gl\") on node \"crc\" DevicePath \"\"" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.350863 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.350874 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2242148-3cf5-4dd7-97f3-ed10f483a2b2-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.908144 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6fdh" event={"ID":"e2242148-3cf5-4dd7-97f3-ed10f483a2b2","Type":"ContainerDied","Data":"df111622a8ff23db77f63dc092608177b775863836966edfa9217c9ed07b3f35"} Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.908191 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6fdh" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.908227 4750 scope.go:117] "RemoveContainer" containerID="e89acd048bf50aa7970c0550c479a1a0440bf666115ca81d16ff2e803eef92f1" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.939583 4750 scope.go:117] "RemoveContainer" containerID="4320354f4424b6c104225c1be7d144c0432544e0477776a21c781a1b1836621d" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.944531 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t6fdh"] Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.950483 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-q9qqf" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.956881 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t6fdh"] Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.958919 4750 scope.go:117] "RemoveContainer" containerID="5f04f21caadb790c47c5c75276e75741958b35675d3d285693ad5bc09fd259af" Mar 09 18:43:18 crc kubenswrapper[4750]: I0309 18:43:18.983158 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-spjm2" Mar 09 18:43:19 crc kubenswrapper[4750]: I0309 18:43:19.383861 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" path="/var/lib/kubelet/pods/e2242148-3cf5-4dd7-97f3-ed10f483a2b2/volumes" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.166566 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt"] Mar 09 18:43:20 crc kubenswrapper[4750]: E0309 18:43:20.166930 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="registry-server" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.166947 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="registry-server" Mar 09 18:43:20 crc kubenswrapper[4750]: E0309 18:43:20.166959 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="extract-content" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.166965 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="extract-content" Mar 09 18:43:20 crc kubenswrapper[4750]: E0309 18:43:20.166976 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="extract-utilities" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.166983 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="extract-utilities" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.167135 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2242148-3cf5-4dd7-97f3-ed10f483a2b2" containerName="registry-server" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.168235 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.171498 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l46j8" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.196974 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt"] Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.282672 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-util\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.282844 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-bundle\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.283022 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7286h\" (UniqueName: \"kubernetes.io/projected/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-kube-api-access-7286h\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.384260 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7286h\" (UniqueName: \"kubernetes.io/projected/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-kube-api-access-7286h\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.384404 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-util\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.384490 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-bundle\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.385295 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-bundle\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.385477 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-util\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.412046 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7286h\" (UniqueName: \"kubernetes.io/projected/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-kube-api-access-7286h\") pod \"5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.501245 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.898288 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt"] Mar 09 18:43:20 crc kubenswrapper[4750]: I0309 18:43:20.927304 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" event={"ID":"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d","Type":"ContainerStarted","Data":"9f6119e63958f26a7c1acf4372688939ec54fcf8f812e3caa6f4db1b48c3c28d"} Mar 09 18:43:21 crc kubenswrapper[4750]: I0309 18:43:21.744266 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:43:21 crc kubenswrapper[4750]: I0309 18:43:21.744767 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:43:21 crc kubenswrapper[4750]: I0309 18:43:21.937933 4750 generic.go:334] "Generic (PLEG): container finished" podID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerID="c5f5da01b3b3b718d4bb2c34ee92a82e08934266a72aebc985995a89aa875b14" exitCode=0 Mar 09 18:43:21 crc kubenswrapper[4750]: I0309 18:43:21.938000 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" event={"ID":"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d","Type":"ContainerDied","Data":"c5f5da01b3b3b718d4bb2c34ee92a82e08934266a72aebc985995a89aa875b14"} Mar 09 18:43:22 crc kubenswrapper[4750]: I0309 18:43:22.949745 4750 generic.go:334] "Generic (PLEG): container finished" podID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerID="35ea8e556410fdd97cc3e51e4f7556f624fb4fb82facb1471a7a07e2c35ffb43" exitCode=0 Mar 09 18:43:22 crc kubenswrapper[4750]: I0309 18:43:22.949799 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" event={"ID":"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d","Type":"ContainerDied","Data":"35ea8e556410fdd97cc3e51e4f7556f624fb4fb82facb1471a7a07e2c35ffb43"} Mar 09 18:43:23 crc kubenswrapper[4750]: I0309 18:43:23.962736 4750 generic.go:334] "Generic (PLEG): container finished" podID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerID="53ea7b543468124abfb2edf24cbb7b2e74a291e22d3b3d2223da1507a032deab" exitCode=0 Mar 09 18:43:23 crc kubenswrapper[4750]: I0309 18:43:23.962828 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" event={"ID":"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d","Type":"ContainerDied","Data":"53ea7b543468124abfb2edf24cbb7b2e74a291e22d3b3d2223da1507a032deab"} Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.248114 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.259055 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-util\") pod \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.259173 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7286h\" (UniqueName: \"kubernetes.io/projected/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-kube-api-access-7286h\") pod \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.259258 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-bundle\") pod \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\" (UID: \"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d\") " Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.260216 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-bundle" (OuterVolumeSpecName: "bundle") pod "f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" (UID: "f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.270899 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-kube-api-access-7286h" (OuterVolumeSpecName: "kube-api-access-7286h") pod "f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" (UID: "f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d"). InnerVolumeSpecName "kube-api-access-7286h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.276875 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-util" (OuterVolumeSpecName: "util") pod "f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" (UID: "f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.360820 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.360856 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-util\") on node \"crc\" DevicePath \"\"" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.360867 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7286h\" (UniqueName: \"kubernetes.io/projected/f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d-kube-api-access-7286h\") on node \"crc\" DevicePath \"\"" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.980190 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" event={"ID":"f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d","Type":"ContainerDied","Data":"9f6119e63958f26a7c1acf4372688939ec54fcf8f812e3caa6f4db1b48c3c28d"} Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.980236 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f6119e63958f26a7c1acf4372688939ec54fcf8f812e3caa6f4db1b48c3c28d" Mar 09 18:43:25 crc kubenswrapper[4750]: I0309 18:43:25.980242 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.529440 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4"] Mar 09 18:43:31 crc kubenswrapper[4750]: E0309 18:43:31.530181 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerName="util" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.530206 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerName="util" Mar 09 18:43:31 crc kubenswrapper[4750]: E0309 18:43:31.530267 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerName="pull" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.530276 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerName="pull" Mar 09 18:43:31 crc kubenswrapper[4750]: E0309 18:43:31.530294 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerName="extract" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.530303 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerName="extract" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.530491 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d" containerName="extract" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.531029 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.533683 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-h6lkr" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.559436 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4"] Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.560162 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65gmw\" (UniqueName: \"kubernetes.io/projected/7fdb116b-f1d9-4e68-ad54-ed2217b6f001-kube-api-access-65gmw\") pod \"openstack-operator-controller-init-787cf98cf6-rb7j4\" (UID: \"7fdb116b-f1d9-4e68-ad54-ed2217b6f001\") " pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.661565 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65gmw\" (UniqueName: \"kubernetes.io/projected/7fdb116b-f1d9-4e68-ad54-ed2217b6f001-kube-api-access-65gmw\") pod \"openstack-operator-controller-init-787cf98cf6-rb7j4\" (UID: \"7fdb116b-f1d9-4e68-ad54-ed2217b6f001\") " pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.684616 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65gmw\" (UniqueName: \"kubernetes.io/projected/7fdb116b-f1d9-4e68-ad54-ed2217b6f001-kube-api-access-65gmw\") pod \"openstack-operator-controller-init-787cf98cf6-rb7j4\" (UID: \"7fdb116b-f1d9-4e68-ad54-ed2217b6f001\") " pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" Mar 09 18:43:31 crc kubenswrapper[4750]: I0309 18:43:31.847298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" Mar 09 18:43:32 crc kubenswrapper[4750]: I0309 18:43:32.090957 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4"] Mar 09 18:43:33 crc kubenswrapper[4750]: I0309 18:43:33.051208 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" event={"ID":"7fdb116b-f1d9-4e68-ad54-ed2217b6f001","Type":"ContainerStarted","Data":"eadc2444ee3198ed4c4c891abf1b3bb11ec5c8e53ebef725129bd8b114067ef8"} Mar 09 18:43:36 crc kubenswrapper[4750]: I0309 18:43:36.079092 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" event={"ID":"7fdb116b-f1d9-4e68-ad54-ed2217b6f001","Type":"ContainerStarted","Data":"3f37addaefa117818217729793c6e2694c174fd5e61efdc09011859b77a92916"} Mar 09 18:43:36 crc kubenswrapper[4750]: I0309 18:43:36.079705 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" Mar 09 18:43:36 crc kubenswrapper[4750]: I0309 18:43:36.131382 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" podStartSLOduration=1.484860188 podStartE2EDuration="5.13135931s" podCreationTimestamp="2026-03-09 18:43:31 +0000 UTC" firstStartedPulling="2026-03-09 18:43:32.097467443 +0000 UTC m=+1093.439939841" lastFinishedPulling="2026-03-09 18:43:35.743966565 +0000 UTC m=+1097.086438963" observedRunningTime="2026-03-09 18:43:36.128722579 +0000 UTC m=+1097.471194977" watchObservedRunningTime="2026-03-09 18:43:36.13135931 +0000 UTC m=+1097.473831728" Mar 09 18:43:41 crc kubenswrapper[4750]: I0309 18:43:41.849842 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-787cf98cf6-rb7j4" Mar 09 18:43:51 crc kubenswrapper[4750]: I0309 18:43:51.744074 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:43:51 crc kubenswrapper[4750]: I0309 18:43:51.744452 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:43:51 crc kubenswrapper[4750]: I0309 18:43:51.744516 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:43:51 crc kubenswrapper[4750]: I0309 18:43:51.745362 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"998a8326795d648e01a6e195fe1e013eaedd975e8692a77e8f1a9a6dca44107f"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:43:51 crc kubenswrapper[4750]: I0309 18:43:51.745429 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://998a8326795d648e01a6e195fe1e013eaedd975e8692a77e8f1a9a6dca44107f" gracePeriod=600 Mar 09 18:43:52 crc kubenswrapper[4750]: I0309 18:43:52.209294 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="998a8326795d648e01a6e195fe1e013eaedd975e8692a77e8f1a9a6dca44107f" exitCode=0 Mar 09 18:43:52 crc kubenswrapper[4750]: I0309 18:43:52.209350 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"998a8326795d648e01a6e195fe1e013eaedd975e8692a77e8f1a9a6dca44107f"} Mar 09 18:43:52 crc kubenswrapper[4750]: I0309 18:43:52.209388 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"aac3b01863d2a16c0edcf63137329635f54035a4884d4a4ea6dddfcacb5b31f0"} Mar 09 18:43:52 crc kubenswrapper[4750]: I0309 18:43:52.209409 4750 scope.go:117] "RemoveContainer" containerID="d522a3b191af7d8f923266bd379a6ce4a390d5d0273a9db4e49eaf7ceef3304f" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.128074 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551364-gv2jc"] Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.129771 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.133114 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.133357 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.134412 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.135391 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551364-gv2jc"] Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.217404 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shnwm\" (UniqueName: \"kubernetes.io/projected/39b5ffd7-804d-47a4-bc5b-f000f7a425a5-kube-api-access-shnwm\") pod \"auto-csr-approver-29551364-gv2jc\" (UID: \"39b5ffd7-804d-47a4-bc5b-f000f7a425a5\") " pod="openshift-infra/auto-csr-approver-29551364-gv2jc" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.318945 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shnwm\" (UniqueName: \"kubernetes.io/projected/39b5ffd7-804d-47a4-bc5b-f000f7a425a5-kube-api-access-shnwm\") pod \"auto-csr-approver-29551364-gv2jc\" (UID: \"39b5ffd7-804d-47a4-bc5b-f000f7a425a5\") " pod="openshift-infra/auto-csr-approver-29551364-gv2jc" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.338073 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shnwm\" (UniqueName: \"kubernetes.io/projected/39b5ffd7-804d-47a4-bc5b-f000f7a425a5-kube-api-access-shnwm\") pod \"auto-csr-approver-29551364-gv2jc\" (UID: \"39b5ffd7-804d-47a4-bc5b-f000f7a425a5\") " pod="openshift-infra/auto-csr-approver-29551364-gv2jc" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.448443 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" Mar 09 18:44:00 crc kubenswrapper[4750]: I0309 18:44:00.749222 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551364-gv2jc"] Mar 09 18:44:01 crc kubenswrapper[4750]: I0309 18:44:01.289873 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" event={"ID":"39b5ffd7-804d-47a4-bc5b-f000f7a425a5","Type":"ContainerStarted","Data":"cbb78a283e8345ae350fd9a903e14580bbe6d688a5a42f8b1a3e4892eaa76b63"} Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.239739 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.240981 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.244913 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-v2mbv" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.247590 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.248767 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.252280 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79h8\" (UniqueName: \"kubernetes.io/projected/607af51a-435c-4f0a-99c2-12c3cf6b2213-kube-api-access-j79h8\") pod \"cinder-operator-controller-manager-55d77d7b5c-86r5x\" (UID: \"607af51a-435c-4f0a-99c2-12c3cf6b2213\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.252375 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftnz2\" (UniqueName: \"kubernetes.io/projected/9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9-kube-api-access-ftnz2\") pod \"barbican-operator-controller-manager-6db6876945-qdwbn\" (UID: \"9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.255501 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-rskz6" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.261722 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.274765 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.279508 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.280342 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.284539 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-sxdj4" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.295523 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.299147 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" event={"ID":"39b5ffd7-804d-47a4-bc5b-f000f7a425a5","Type":"ContainerStarted","Data":"a1e7416aecca045ff84400c55306c667a0c4ff18352fdd539eb0c924c2c24b47"} Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.311374 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.312458 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.323007 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-gfg2c" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.327386 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.328501 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.337299 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-wrmfw" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.344001 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.353314 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbgbf\" (UniqueName: \"kubernetes.io/projected/bf342fef-8819-4328-9551-13d21041bff6-kube-api-access-qbgbf\") pod \"designate-operator-controller-manager-5d87c9d997-czkhf\" (UID: \"bf342fef-8819-4328-9551-13d21041bff6\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.353367 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftnz2\" (UniqueName: \"kubernetes.io/projected/9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9-kube-api-access-ftnz2\") pod \"barbican-operator-controller-manager-6db6876945-qdwbn\" (UID: \"9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.353414 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mcc8\" (UniqueName: \"kubernetes.io/projected/434cb9b7-d90c-4081-8232-004bbd4cb8b2-kube-api-access-8mcc8\") pod \"heat-operator-controller-manager-cf99c678f-xrjb7\" (UID: \"434cb9b7-d90c-4081-8232-004bbd4cb8b2\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.353437 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s65n\" (UniqueName: \"kubernetes.io/projected/c7f102b3-0989-4582-91d1-49a22c261371-kube-api-access-8s65n\") pod \"glance-operator-controller-manager-64db6967f8-wl9k2\" (UID: \"c7f102b3-0989-4582-91d1-49a22c261371\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.353518 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79h8\" (UniqueName: \"kubernetes.io/projected/607af51a-435c-4f0a-99c2-12c3cf6b2213-kube-api-access-j79h8\") pod \"cinder-operator-controller-manager-55d77d7b5c-86r5x\" (UID: \"607af51a-435c-4f0a-99c2-12c3cf6b2213\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.362546 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.366235 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.371820 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mdf6m" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.382928 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftnz2\" (UniqueName: \"kubernetes.io/projected/9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9-kube-api-access-ftnz2\") pod \"barbican-operator-controller-manager-6db6876945-qdwbn\" (UID: \"9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.384439 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.389260 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.390800 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79h8\" (UniqueName: \"kubernetes.io/projected/607af51a-435c-4f0a-99c2-12c3cf6b2213-kube-api-access-j79h8\") pod \"cinder-operator-controller-manager-55d77d7b5c-86r5x\" (UID: \"607af51a-435c-4f0a-99c2-12c3cf6b2213\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.416300 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.417367 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.426405 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gpjhk" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.426584 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.430436 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" podStartSLOduration=1.514932089 podStartE2EDuration="2.430420101s" podCreationTimestamp="2026-03-09 18:44:00 +0000 UTC" firstStartedPulling="2026-03-09 18:44:00.763525836 +0000 UTC m=+1122.105998234" lastFinishedPulling="2026-03-09 18:44:01.679013828 +0000 UTC m=+1123.021486246" observedRunningTime="2026-03-09 18:44:02.383318178 +0000 UTC m=+1123.725790576" watchObservedRunningTime="2026-03-09 18:44:02.430420101 +0000 UTC m=+1123.772892499" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.434222 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.450864 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.452043 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.456081 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qnwl\" (UniqueName: \"kubernetes.io/projected/6f320add-8d8a-49c6-a8bd-c286772ee907-kube-api-access-8qnwl\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.456271 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7x4g\" (UniqueName: \"kubernetes.io/projected/f4e026ba-f7dd-4c14-b09e-e00e6cfae20d-kube-api-access-t7x4g\") pod \"horizon-operator-controller-manager-78bc7f9bd9-tkcr6\" (UID: \"f4e026ba-f7dd-4c14-b09e-e00e6cfae20d\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.456374 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq46w\" (UniqueName: \"kubernetes.io/projected/3f316fad-80d2-46fb-ab97-d3f2b0559387-kube-api-access-xq46w\") pod \"ironic-operator-controller-manager-545456dc4-lmj58\" (UID: \"3f316fad-80d2-46fb-ab97-d3f2b0559387\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.456443 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.456541 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbgbf\" (UniqueName: \"kubernetes.io/projected/bf342fef-8819-4328-9551-13d21041bff6-kube-api-access-qbgbf\") pod \"designate-operator-controller-manager-5d87c9d997-czkhf\" (UID: \"bf342fef-8819-4328-9551-13d21041bff6\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.456668 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mcc8\" (UniqueName: \"kubernetes.io/projected/434cb9b7-d90c-4081-8232-004bbd4cb8b2-kube-api-access-8mcc8\") pod \"heat-operator-controller-manager-cf99c678f-xrjb7\" (UID: \"434cb9b7-d90c-4081-8232-004bbd4cb8b2\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.456770 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s65n\" (UniqueName: \"kubernetes.io/projected/c7f102b3-0989-4582-91d1-49a22c261371-kube-api-access-8s65n\") pod \"glance-operator-controller-manager-64db6967f8-wl9k2\" (UID: \"c7f102b3-0989-4582-91d1-49a22c261371\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.461756 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-pnd76" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.478672 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbgbf\" (UniqueName: \"kubernetes.io/projected/bf342fef-8819-4328-9551-13d21041bff6-kube-api-access-qbgbf\") pod \"designate-operator-controller-manager-5d87c9d997-czkhf\" (UID: \"bf342fef-8819-4328-9551-13d21041bff6\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.478917 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.491835 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.492593 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.494603 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s65n\" (UniqueName: \"kubernetes.io/projected/c7f102b3-0989-4582-91d1-49a22c261371-kube-api-access-8s65n\") pod \"glance-operator-controller-manager-64db6967f8-wl9k2\" (UID: \"c7f102b3-0989-4582-91d1-49a22c261371\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.496921 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-jpgz7" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.497789 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mcc8\" (UniqueName: \"kubernetes.io/projected/434cb9b7-d90c-4081-8232-004bbd4cb8b2-kube-api-access-8mcc8\") pod \"heat-operator-controller-manager-cf99c678f-xrjb7\" (UID: \"434cb9b7-d90c-4081-8232-004bbd4cb8b2\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.512870 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.523194 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.524015 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.528224 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-zjvmf" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.538299 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.539321 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.550038 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-b75nk" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.554244 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.562726 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.564340 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6lvd\" (UniqueName: \"kubernetes.io/projected/e671bb3b-ca69-4229-9e9b-e823bfcf2f92-kube-api-access-h6lvd\") pod \"manila-operator-controller-manager-67d996989d-wrkj4\" (UID: \"e671bb3b-ca69-4229-9e9b-e823bfcf2f92\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.564463 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7x4g\" (UniqueName: \"kubernetes.io/projected/f4e026ba-f7dd-4c14-b09e-e00e6cfae20d-kube-api-access-t7x4g\") pod \"horizon-operator-controller-manager-78bc7f9bd9-tkcr6\" (UID: \"f4e026ba-f7dd-4c14-b09e-e00e6cfae20d\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.564555 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq46w\" (UniqueName: \"kubernetes.io/projected/3f316fad-80d2-46fb-ab97-d3f2b0559387-kube-api-access-xq46w\") pod \"ironic-operator-controller-manager-545456dc4-lmj58\" (UID: \"3f316fad-80d2-46fb-ab97-d3f2b0559387\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.564648 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.564752 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9gh5\" (UniqueName: \"kubernetes.io/projected/d8208475-d487-47ee-8f03-fa4379d3c917-kube-api-access-f9gh5\") pod \"mariadb-operator-controller-manager-7b6bfb6475-znrqs\" (UID: \"d8208475-d487-47ee-8f03-fa4379d3c917\") " pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.564906 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qnwl\" (UniqueName: \"kubernetes.io/projected/6f320add-8d8a-49c6-a8bd-c286772ee907-kube-api-access-8qnwl\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.565282 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47l5v\" (UniqueName: \"kubernetes.io/projected/a4506343-59e4-4442-8894-fdb158c82316-kube-api-access-47l5v\") pod \"keystone-operator-controller-manager-7c789f89c6-j799f\" (UID: \"a4506343-59e4-4442-8894-fdb158c82316\") " pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" Mar 09 18:44:02 crc kubenswrapper[4750]: E0309 18:44:02.565467 4750 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:02 crc kubenswrapper[4750]: E0309 18:44:02.565570 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert podName:6f320add-8d8a-49c6-a8bd-c286772ee907 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:03.065552363 +0000 UTC m=+1124.408024771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert") pod "infra-operator-controller-manager-f7fcc58b9-8pvfn" (UID: "6f320add-8d8a-49c6-a8bd-c286772ee907") : secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.571768 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.572699 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.573749 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.574691 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.587060 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-c672w" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.599684 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.633389 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7x4g\" (UniqueName: \"kubernetes.io/projected/f4e026ba-f7dd-4c14-b09e-e00e6cfae20d-kube-api-access-t7x4g\") pod \"horizon-operator-controller-manager-78bc7f9bd9-tkcr6\" (UID: \"f4e026ba-f7dd-4c14-b09e-e00e6cfae20d\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.633544 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qnwl\" (UniqueName: \"kubernetes.io/projected/6f320add-8d8a-49c6-a8bd-c286772ee907-kube-api-access-8qnwl\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.634381 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq46w\" (UniqueName: \"kubernetes.io/projected/3f316fad-80d2-46fb-ab97-d3f2b0559387-kube-api-access-xq46w\") pod \"ironic-operator-controller-manager-545456dc4-lmj58\" (UID: \"3f316fad-80d2-46fb-ab97-d3f2b0559387\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.637376 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.648767 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.684748 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.685413 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6lvd\" (UniqueName: \"kubernetes.io/projected/e671bb3b-ca69-4229-9e9b-e823bfcf2f92-kube-api-access-h6lvd\") pod \"manila-operator-controller-manager-67d996989d-wrkj4\" (UID: \"e671bb3b-ca69-4229-9e9b-e823bfcf2f92\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.685555 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9gh5\" (UniqueName: \"kubernetes.io/projected/d8208475-d487-47ee-8f03-fa4379d3c917-kube-api-access-f9gh5\") pod \"mariadb-operator-controller-manager-7b6bfb6475-znrqs\" (UID: \"d8208475-d487-47ee-8f03-fa4379d3c917\") " pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.685643 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47l5v\" (UniqueName: \"kubernetes.io/projected/a4506343-59e4-4442-8894-fdb158c82316-kube-api-access-47l5v\") pod \"keystone-operator-controller-manager-7c789f89c6-j799f\" (UID: \"a4506343-59e4-4442-8894-fdb158c82316\") " pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.685690 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zclp\" (UniqueName: \"kubernetes.io/projected/1dc593d7-29a0-4a18-aecc-81f22e19e9b9-kube-api-access-9zclp\") pod \"neutron-operator-controller-manager-54688575f-n8ldk\" (UID: \"1dc593d7-29a0-4a18-aecc-81f22e19e9b9\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.711856 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6lvd\" (UniqueName: \"kubernetes.io/projected/e671bb3b-ca69-4229-9e9b-e823bfcf2f92-kube-api-access-h6lvd\") pod \"manila-operator-controller-manager-67d996989d-wrkj4\" (UID: \"e671bb3b-ca69-4229-9e9b-e823bfcf2f92\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.734240 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.735363 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.748653 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47l5v\" (UniqueName: \"kubernetes.io/projected/a4506343-59e4-4442-8894-fdb158c82316-kube-api-access-47l5v\") pod \"keystone-operator-controller-manager-7c789f89c6-j799f\" (UID: \"a4506343-59e4-4442-8894-fdb158c82316\") " pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.754163 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-zf2g2" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.769729 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.783516 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9gh5\" (UniqueName: \"kubernetes.io/projected/d8208475-d487-47ee-8f03-fa4379d3c917-kube-api-access-f9gh5\") pod \"mariadb-operator-controller-manager-7b6bfb6475-znrqs\" (UID: \"d8208475-d487-47ee-8f03-fa4379d3c917\") " pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.783598 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.787992 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zclp\" (UniqueName: \"kubernetes.io/projected/1dc593d7-29a0-4a18-aecc-81f22e19e9b9-kube-api-access-9zclp\") pod \"neutron-operator-controller-manager-54688575f-n8ldk\" (UID: \"1dc593d7-29a0-4a18-aecc-81f22e19e9b9\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.804695 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.806391 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.809068 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-pp8qg" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.810307 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.812773 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zclp\" (UniqueName: \"kubernetes.io/projected/1dc593d7-29a0-4a18-aecc-81f22e19e9b9-kube-api-access-9zclp\") pod \"neutron-operator-controller-manager-54688575f-n8ldk\" (UID: \"1dc593d7-29a0-4a18-aecc-81f22e19e9b9\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.825399 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.826691 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.830518 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-4sldm" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.848165 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.849503 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.851962 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qfh5h" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.852414 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.862027 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.865674 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.879988 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.880674 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hmj8h" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.891269 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.893056 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsrq8\" (UniqueName: \"kubernetes.io/projected/14f86f92-1676-44e4-a56c-e611b59afe30-kube-api-access-fsrq8\") pod \"nova-operator-controller-manager-74b6b5dc96-qp844\" (UID: \"14f86f92-1676-44e4-a56c-e611b59afe30\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.893132 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72l5t\" (UniqueName: \"kubernetes.io/projected/c42f78da-7508-4f90-bee6-c05e5cbd0f86-kube-api-access-72l5t\") pod \"octavia-operator-controller-manager-5d86c7ddb7-dnt7g\" (UID: \"c42f78da-7508-4f90-bee6-c05e5cbd0f86\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.896532 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.903034 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.907855 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-9x8ns" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.935732 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.939824 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms"] Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.940813 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" Mar 09 18:44:02 crc kubenswrapper[4750]: I0309 18:44:02.954685 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-qgwnf" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:02.994851 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2hdp\" (UniqueName: \"kubernetes.io/projected/ccf58e0b-b6b3-443a-b215-087bb55da05c-kube-api-access-s2hdp\") pod \"swift-operator-controller-manager-9b9ff9f4d-skdbr\" (UID: \"ccf58e0b-b6b3-443a-b215-087bb55da05c\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:02.994919 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz968\" (UniqueName: \"kubernetes.io/projected/1507d1a6-11a2-4a35-9c3c-1c360e4c0b32-kube-api-access-hz968\") pod \"ovn-operator-controller-manager-75684d597f-hj5hw\" (UID: \"1507d1a6-11a2-4a35-9c3c-1c360e4c0b32\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:02.994957 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj59l\" (UniqueName: \"kubernetes.io/projected/93ac04dd-4b80-4e20-9113-eadef2054d0e-kube-api-access-pj59l\") pod \"placement-operator-controller-manager-648564c9fc-2l9js\" (UID: \"93ac04dd-4b80-4e20-9113-eadef2054d0e\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:02.994975 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:02.995000 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsrq8\" (UniqueName: \"kubernetes.io/projected/14f86f92-1676-44e4-a56c-e611b59afe30-kube-api-access-fsrq8\") pod \"nova-operator-controller-manager-74b6b5dc96-qp844\" (UID: \"14f86f92-1676-44e4-a56c-e611b59afe30\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:02.995041 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72l5t\" (UniqueName: \"kubernetes.io/projected/c42f78da-7508-4f90-bee6-c05e5cbd0f86-kube-api-access-72l5t\") pod \"octavia-operator-controller-manager-5d86c7ddb7-dnt7g\" (UID: \"c42f78da-7508-4f90-bee6-c05e5cbd0f86\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:02.995087 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdxwx\" (UniqueName: \"kubernetes.io/projected/89799193-9df0-4505-8116-d49a86d33add-kube-api-access-pdxwx\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.002695 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.007432 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.007598 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.023249 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.041148 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsrq8\" (UniqueName: \"kubernetes.io/projected/14f86f92-1676-44e4-a56c-e611b59afe30-kube-api-access-fsrq8\") pod \"nova-operator-controller-manager-74b6b5dc96-qp844\" (UID: \"14f86f92-1676-44e4-a56c-e611b59afe30\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.041693 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72l5t\" (UniqueName: \"kubernetes.io/projected/c42f78da-7508-4f90-bee6-c05e5cbd0f86-kube-api-access-72l5t\") pod \"octavia-operator-controller-manager-5d86c7ddb7-dnt7g\" (UID: \"c42f78da-7508-4f90-bee6-c05e5cbd0f86\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.050557 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.066079 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.086470 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.094292 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097095 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097272 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdxwx\" (UniqueName: \"kubernetes.io/projected/89799193-9df0-4505-8116-d49a86d33add-kube-api-access-pdxwx\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097387 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2hdp\" (UniqueName: \"kubernetes.io/projected/ccf58e0b-b6b3-443a-b215-087bb55da05c-kube-api-access-s2hdp\") pod \"swift-operator-controller-manager-9b9ff9f4d-skdbr\" (UID: \"ccf58e0b-b6b3-443a-b215-087bb55da05c\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097471 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz968\" (UniqueName: \"kubernetes.io/projected/1507d1a6-11a2-4a35-9c3c-1c360e4c0b32-kube-api-access-hz968\") pod \"ovn-operator-controller-manager-75684d597f-hj5hw\" (UID: \"1507d1a6-11a2-4a35-9c3c-1c360e4c0b32\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097550 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccvmw\" (UniqueName: \"kubernetes.io/projected/e403dea7-17d0-40db-9b29-e50f33175503-kube-api-access-ccvmw\") pod \"telemetry-operator-controller-manager-5fdb694969-rvpms\" (UID: \"e403dea7-17d0-40db-9b29-e50f33175503\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097677 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj59l\" (UniqueName: \"kubernetes.io/projected/93ac04dd-4b80-4e20-9113-eadef2054d0e-kube-api-access-pj59l\") pod \"placement-operator-controller-manager-648564c9fc-2l9js\" (UID: \"93ac04dd-4b80-4e20-9113-eadef2054d0e\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097755 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.097984 4750 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.097985 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.098095 4750 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.098106 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert podName:89799193-9df0-4505-8116-d49a86d33add nodeName:}" failed. No retries permitted until 2026-03-09 18:44:03.598083689 +0000 UTC m=+1124.940556087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" (UID: "89799193-9df0-4505-8116-d49a86d33add") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.098272 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert podName:6f320add-8d8a-49c6-a8bd-c286772ee907 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:04.098238093 +0000 UTC m=+1125.440710591 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert") pod "infra-operator-controller-manager-f7fcc58b9-8pvfn" (UID: "6f320add-8d8a-49c6-a8bd-c286772ee907") : secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.104705 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.139564 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-gxfdl" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.147535 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj59l\" (UniqueName: \"kubernetes.io/projected/93ac04dd-4b80-4e20-9113-eadef2054d0e-kube-api-access-pj59l\") pod \"placement-operator-controller-manager-648564c9fc-2l9js\" (UID: \"93ac04dd-4b80-4e20-9113-eadef2054d0e\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.147930 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.149151 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz968\" (UniqueName: \"kubernetes.io/projected/1507d1a6-11a2-4a35-9c3c-1c360e4c0b32-kube-api-access-hz968\") pod \"ovn-operator-controller-manager-75684d597f-hj5hw\" (UID: \"1507d1a6-11a2-4a35-9c3c-1c360e4c0b32\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.158937 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdxwx\" (UniqueName: \"kubernetes.io/projected/89799193-9df0-4505-8116-d49a86d33add-kube-api-access-pdxwx\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.160159 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.161705 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2hdp\" (UniqueName: \"kubernetes.io/projected/ccf58e0b-b6b3-443a-b215-087bb55da05c-kube-api-access-s2hdp\") pod \"swift-operator-controller-manager-9b9ff9f4d-skdbr\" (UID: \"ccf58e0b-b6b3-443a-b215-087bb55da05c\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.201136 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.203197 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qvb9\" (UniqueName: \"kubernetes.io/projected/8c687b21-5077-4d33-af70-07b01956d904-kube-api-access-7qvb9\") pod \"test-operator-controller-manager-55b5ff4dbb-z96st\" (UID: \"8c687b21-5077-4d33-af70-07b01956d904\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.203282 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccvmw\" (UniqueName: \"kubernetes.io/projected/e403dea7-17d0-40db-9b29-e50f33175503-kube-api-access-ccvmw\") pod \"telemetry-operator-controller-manager-5fdb694969-rvpms\" (UID: \"e403dea7-17d0-40db-9b29-e50f33175503\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.206531 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.207803 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.211080 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lmjzg" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.211998 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.225533 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccvmw\" (UniqueName: \"kubernetes.io/projected/e403dea7-17d0-40db-9b29-e50f33175503-kube-api-access-ccvmw\") pod \"telemetry-operator-controller-manager-5fdb694969-rvpms\" (UID: \"e403dea7-17d0-40db-9b29-e50f33175503\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.259597 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.280768 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.282038 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.284476 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.284683 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.285131 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zc9z6" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.287560 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.316753 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.335342 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.338286 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.352021 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-64rht" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.354283 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzmpf\" (UniqueName: \"kubernetes.io/projected/6bd1943b-5152-4d4e-9fa9-8c2f8ea42072-kube-api-access-hzmpf\") pod \"watcher-operator-controller-manager-668c5c65dc-rt6bl\" (UID: \"6bd1943b-5152-4d4e-9fa9-8c2f8ea42072\") " pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.354420 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qvb9\" (UniqueName: \"kubernetes.io/projected/8c687b21-5077-4d33-af70-07b01956d904-kube-api-access-7qvb9\") pod \"test-operator-controller-manager-55b5ff4dbb-z96st\" (UID: \"8c687b21-5077-4d33-af70-07b01956d904\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.364988 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.387557 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.445510 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qvb9\" (UniqueName: \"kubernetes.io/projected/8c687b21-5077-4d33-af70-07b01956d904-kube-api-access-7qvb9\") pod \"test-operator-controller-manager-55b5ff4dbb-z96st\" (UID: \"8c687b21-5077-4d33-af70-07b01956d904\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.457620 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.457867 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.457938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwg84\" (UniqueName: \"kubernetes.io/projected/a91098d3-ee59-4f84-ad48-ee7613a8c7da-kube-api-access-qwg84\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fbjb2\" (UID: \"a91098d3-ee59-4f84-ad48-ee7613a8c7da\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.458946 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzmpf\" (UniqueName: \"kubernetes.io/projected/6bd1943b-5152-4d4e-9fa9-8c2f8ea42072-kube-api-access-hzmpf\") pod \"watcher-operator-controller-manager-668c5c65dc-rt6bl\" (UID: \"6bd1943b-5152-4d4e-9fa9-8c2f8ea42072\") " pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.461785 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dkrn\" (UniqueName: \"kubernetes.io/projected/874a050d-1e92-4087-a655-b6fdb66a47e9-kube-api-access-9dkrn\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.465013 4750 generic.go:334] "Generic (PLEG): container finished" podID="39b5ffd7-804d-47a4-bc5b-f000f7a425a5" containerID="a1e7416aecca045ff84400c55306c667a0c4ff18352fdd539eb0c924c2c24b47" exitCode=0 Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.467469 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" event={"ID":"39b5ffd7-804d-47a4-bc5b-f000f7a425a5","Type":"ContainerDied","Data":"a1e7416aecca045ff84400c55306c667a0c4ff18352fdd539eb0c924c2c24b47"} Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.510025 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzmpf\" (UniqueName: \"kubernetes.io/projected/6bd1943b-5152-4d4e-9fa9-8c2f8ea42072-kube-api-access-hzmpf\") pod \"watcher-operator-controller-manager-668c5c65dc-rt6bl\" (UID: \"6bd1943b-5152-4d4e-9fa9-8c2f8ea42072\") " pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.531619 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.539820 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.549976 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.552770 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.567747 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dkrn\" (UniqueName: \"kubernetes.io/projected/874a050d-1e92-4087-a655-b6fdb66a47e9-kube-api-access-9dkrn\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.567991 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.568050 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.568096 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwg84\" (UniqueName: \"kubernetes.io/projected/a91098d3-ee59-4f84-ad48-ee7613a8c7da-kube-api-access-qwg84\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fbjb2\" (UID: \"a91098d3-ee59-4f84-ad48-ee7613a8c7da\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.568755 4750 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.568801 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:04.068785237 +0000 UTC m=+1125.411257635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "metrics-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.568936 4750 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.568957 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:04.068950522 +0000 UTC m=+1125.411422920 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.597470 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dkrn\" (UniqueName: \"kubernetes.io/projected/874a050d-1e92-4087-a655-b6fdb66a47e9-kube-api-access-9dkrn\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.599896 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwg84\" (UniqueName: \"kubernetes.io/projected/a91098d3-ee59-4f84-ad48-ee7613a8c7da-kube-api-access-qwg84\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fbjb2\" (UID: \"a91098d3-ee59-4f84-ad48-ee7613a8c7da\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.628673 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.672245 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.672492 4750 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: E0309 18:44:03.672562 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert podName:89799193-9df0-4505-8116-d49a86d33add nodeName:}" failed. No retries permitted until 2026-03-09 18:44:04.672534938 +0000 UTC m=+1126.015007336 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" (UID: "89799193-9df0-4505-8116-d49a86d33add") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.731189 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.860410 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.890189 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.928821 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6"] Mar 09 18:44:03 crc kubenswrapper[4750]: I0309 18:44:03.979033 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.080757 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.082048 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.082124 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.082365 4750 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.082449 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:05.082428737 +0000 UTC m=+1126.424901135 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "webhook-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.082850 4750 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.082888 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:05.082879749 +0000 UTC m=+1126.425352137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "metrics-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.095793 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.105984 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.110816 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs"] Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.128687 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8208475_d487_47ee_8f03_fa4379d3c917.slice/crio-fc072cab58fde8665c88c8a7ef815c1a6a6c609327afc4b6b8fe55eb7a4a8d76 WatchSource:0}: Error finding container fc072cab58fde8665c88c8a7ef815c1a6a6c609327afc4b6b8fe55eb7a4a8d76: Status 404 returned error can't find the container with id fc072cab58fde8665c88c8a7ef815c1a6a6c609327afc4b6b8fe55eb7a4a8d76 Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.135936 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode671bb3b_ca69_4229_9e9b_e823bfcf2f92.slice/crio-409fd4f0601fb43fc34e2fd7a0e36f2b50a32832fd861ca9f3a8e991b97fa2cb WatchSource:0}: Error finding container 409fd4f0601fb43fc34e2fd7a0e36f2b50a32832fd861ca9f3a8e991b97fa2cb: Status 404 returned error can't find the container with id 409fd4f0601fb43fc34e2fd7a0e36f2b50a32832fd861ca9f3a8e991b97fa2cb Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.141810 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1dc593d7_29a0_4a18_aecc_81f22e19e9b9.slice/crio-a75631c7e595bf102afba0ed395f7a71bf9bddd333de1c053a70e3451b9751cb WatchSource:0}: Error finding container a75631c7e595bf102afba0ed395f7a71bf9bddd333de1c053a70e3451b9751cb: Status 404 returned error can't find the container with id a75631c7e595bf102afba0ed395f7a71bf9bddd333de1c053a70e3451b9751cb Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.187391 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.187568 4750 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.187653 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert podName:6f320add-8d8a-49c6-a8bd-c286772ee907 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:06.187617656 +0000 UTC m=+1127.530090064 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert") pod "infra-operator-controller-manager-f7fcc58b9-8pvfn" (UID: "6f320add-8d8a-49c6-a8bd-c286772ee907") : secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.201255 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.344503 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.355572 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.362268 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.369225 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw"] Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.369297 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccf58e0b_b6b3_443a_b215_087bb55da05c.slice/crio-a2db1287eed7cda728f8ceb8edc03ed365fa7603cce4f124d327577c9b03e034 WatchSource:0}: Error finding container a2db1287eed7cda728f8ceb8edc03ed365fa7603cce4f124d327577c9b03e034: Status 404 returned error can't find the container with id a2db1287eed7cda728f8ceb8edc03ed365fa7603cce4f124d327577c9b03e034 Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.371147 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode403dea7_17d0_40db_9b29_e50f33175503.slice/crio-4318b006fe56e9d0b61218d08aa7d9062137bec194604aaff695be594ae6afbd WatchSource:0}: Error finding container 4318b006fe56e9d0b61218d08aa7d9062137bec194604aaff695be594ae6afbd: Status 404 returned error can't find the container with id 4318b006fe56e9d0b61218d08aa7d9062137bec194604aaff695be594ae6afbd Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.382455 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc42f78da_7508_4f90_bee6_c05e5cbd0f86.slice/crio-50262028bedff548d1d176fdfb3efce44e890f26f9d9a9bbf3c52faee5d34aac WatchSource:0}: Error finding container 50262028bedff548d1d176fdfb3efce44e890f26f9d9a9bbf3c52faee5d34aac: Status 404 returned error can't find the container with id 50262028bedff548d1d176fdfb3efce44e890f26f9d9a9bbf3c52faee5d34aac Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.386779 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:2d59045b8d8e6f9c5483c4fdda7c5057218d553200dc4bcf26789980ac1d9abd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-72l5t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5d86c7ddb7-dnt7g_openstack-operators(c42f78da-7508-4f90-bee6-c05e5cbd0f86): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.386778 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:9f73c84a9581b5739d8da333c7b64403d7b7ca284b22c624d0effe07f3d2819c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hz968,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-75684d597f-hj5hw_openstack-operators(1507d1a6-11a2-4a35-9c3c-1c360e4c0b32): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.387990 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" podUID="1507d1a6-11a2-4a35-9c3c-1c360e4c0b32" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.388010 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" podUID="c42f78da-7508-4f90-bee6-c05e5cbd0f86" Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.473773 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" event={"ID":"9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9","Type":"ContainerStarted","Data":"2343d8d9717e451daa9e54ba6c6a1fa870592621844134a720506a165898c270"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.474848 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" event={"ID":"14f86f92-1676-44e4-a56c-e611b59afe30","Type":"ContainerStarted","Data":"2461d6ede006f69a8270ea1b70d97150b2ef6e9f4194251a33a25d6aee28f6f7"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.477545 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" event={"ID":"ccf58e0b-b6b3-443a-b215-087bb55da05c","Type":"ContainerStarted","Data":"a2db1287eed7cda728f8ceb8edc03ed365fa7603cce4f124d327577c9b03e034"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.480603 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" event={"ID":"f4e026ba-f7dd-4c14-b09e-e00e6cfae20d","Type":"ContainerStarted","Data":"4d31e0af1e8cb430e843eea713307d8b7bb680ad657221cc38f17c97a1a9feca"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.488114 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" event={"ID":"1dc593d7-29a0-4a18-aecc-81f22e19e9b9","Type":"ContainerStarted","Data":"a75631c7e595bf102afba0ed395f7a71bf9bddd333de1c053a70e3451b9751cb"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.491449 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" event={"ID":"a4506343-59e4-4442-8894-fdb158c82316","Type":"ContainerStarted","Data":"a2282f312b0453fdfa0996e0258de146918f80d7e37d106fb323ed5445856068"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.494391 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" event={"ID":"1507d1a6-11a2-4a35-9c3c-1c360e4c0b32","Type":"ContainerStarted","Data":"e7dad18a1c706a741b1a4c36c91e7d985e5d79584323cdce16510b83efeda1f7"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.502532 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" event={"ID":"bf342fef-8819-4328-9551-13d21041bff6","Type":"ContainerStarted","Data":"a22b35a3496fe876d5d1646f523f5bfa7e2afd075117bec66da6635eddcfe367"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.506225 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" event={"ID":"607af51a-435c-4f0a-99c2-12c3cf6b2213","Type":"ContainerStarted","Data":"a2de638030741d98d3780e057e41728fa72151d6c9d99797a65aa1e11c177d1d"} Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.507102 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:9f73c84a9581b5739d8da333c7b64403d7b7ca284b22c624d0effe07f3d2819c\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" podUID="1507d1a6-11a2-4a35-9c3c-1c360e4c0b32" Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.509466 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" event={"ID":"d8208475-d487-47ee-8f03-fa4379d3c917","Type":"ContainerStarted","Data":"fc072cab58fde8665c88c8a7ef815c1a6a6c609327afc4b6b8fe55eb7a4a8d76"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.524781 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" event={"ID":"e671bb3b-ca69-4229-9e9b-e823bfcf2f92","Type":"ContainerStarted","Data":"409fd4f0601fb43fc34e2fd7a0e36f2b50a32832fd861ca9f3a8e991b97fa2cb"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.529669 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" event={"ID":"e403dea7-17d0-40db-9b29-e50f33175503","Type":"ContainerStarted","Data":"4318b006fe56e9d0b61218d08aa7d9062137bec194604aaff695be594ae6afbd"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.533481 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" event={"ID":"c42f78da-7508-4f90-bee6-c05e5cbd0f86","Type":"ContainerStarted","Data":"50262028bedff548d1d176fdfb3efce44e890f26f9d9a9bbf3c52faee5d34aac"} Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.541659 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:2d59045b8d8e6f9c5483c4fdda7c5057218d553200dc4bcf26789980ac1d9abd\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" podUID="c42f78da-7508-4f90-bee6-c05e5cbd0f86" Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.545871 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" event={"ID":"c7f102b3-0989-4582-91d1-49a22c261371","Type":"ContainerStarted","Data":"57ac3338eced9b9710cb6884fac0837668a2c67aeaa1f486d978ab924395247c"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.565007 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.565597 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" event={"ID":"3f316fad-80d2-46fb-ab97-d3f2b0559387","Type":"ContainerStarted","Data":"39a493385628d99f7ab19eb702aaa059d9794e318ac9cab3c221da169751998f"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.566965 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" event={"ID":"434cb9b7-d90c-4081-8232-004bbd4cb8b2","Type":"ContainerStarted","Data":"2aa1d4865546a239fb0417876bf6c4c519409abf860b4ea588b2544fec5ed0cc"} Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.571123 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js"] Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.588742 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl"] Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.592104 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bd1943b_5152_4d4e_9fa9_8c2f8ea42072.slice/crio-a3775aa1917dc4cfbb9795a342a67ae57ac0a29a52fde1dedeede0cd3c454815 WatchSource:0}: Error finding container a3775aa1917dc4cfbb9795a342a67ae57ac0a29a52fde1dedeede0cd3c454815: Status 404 returned error can't find the container with id a3775aa1917dc4cfbb9795a342a67ae57ac0a29a52fde1dedeede0cd3c454815 Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.592475 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93ac04dd_4b80_4e20_9113_eadef2054d0e.slice/crio-238a0f1fc915760ded8cd11ce87d0ada51c067df8e90278ba385274c47141e33 WatchSource:0}: Error finding container 238a0f1fc915760ded8cd11ce87d0ada51c067df8e90278ba385274c47141e33: Status 404 returned error can't find the container with id 238a0f1fc915760ded8cd11ce87d0ada51c067df8e90278ba385274c47141e33 Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.595847 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:bb939885bd04593ad03af901adb77ee2a2d18529b328c23288c7cc7a2ba5282e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pj59l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-648564c9fc-2l9js_openstack-operators(93ac04dd-4b80-4e20-9113-eadef2054d0e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.595901 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda91098d3_ee59_4f84_ad48_ee7613a8c7da.slice/crio-17cd4cb1e8141d0ddc272304c62df3271871d501eaf97d67a2eec679c47fef51 WatchSource:0}: Error finding container 17cd4cb1e8141d0ddc272304c62df3271871d501eaf97d67a2eec679c47fef51: Status 404 returned error can't find the container with id 17cd4cb1e8141d0ddc272304c62df3271871d501eaf97d67a2eec679c47fef51 Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.597309 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" podUID="93ac04dd-4b80-4e20-9113-eadef2054d0e" Mar 09 18:44:04 crc kubenswrapper[4750]: W0309 18:44:04.598168 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c687b21_5077_4d33_af70_07b01956d904.slice/crio-d05340dd53dfeabb28632ab7f89daf2587a13470985c4a72881c83ce66af0d7b WatchSource:0}: Error finding container d05340dd53dfeabb28632ab7f89daf2587a13470985c4a72881c83ce66af0d7b: Status 404 returned error can't find the container with id d05340dd53dfeabb28632ab7f89daf2587a13470985c4a72881c83ce66af0d7b Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.600830 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2"] Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.602608 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7qvb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-55b5ff4dbb-z96st_openstack-operators(8c687b21-5077-4d33-af70-07b01956d904): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.603774 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" podUID="8c687b21-5077-4d33-af70-07b01956d904" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.606269 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qwg84,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-fbjb2_openstack-operators(a91098d3-ee59-4f84-ad48-ee7613a8c7da): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.608044 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" podUID="a91098d3-ee59-4f84-ad48-ee7613a8c7da" Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.705528 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.705947 4750 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: E0309 18:44:04.706069 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert podName:89799193-9df0-4505-8116-d49a86d33add nodeName:}" failed. No retries permitted until 2026-03-09 18:44:06.706023583 +0000 UTC m=+1128.048495981 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" (UID: "89799193-9df0-4505-8116-d49a86d33add") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:04 crc kubenswrapper[4750]: I0309 18:44:04.910654 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.013927 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shnwm\" (UniqueName: \"kubernetes.io/projected/39b5ffd7-804d-47a4-bc5b-f000f7a425a5-kube-api-access-shnwm\") pod \"39b5ffd7-804d-47a4-bc5b-f000f7a425a5\" (UID: \"39b5ffd7-804d-47a4-bc5b-f000f7a425a5\") " Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.021316 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39b5ffd7-804d-47a4-bc5b-f000f7a425a5-kube-api-access-shnwm" (OuterVolumeSpecName: "kube-api-access-shnwm") pod "39b5ffd7-804d-47a4-bc5b-f000f7a425a5" (UID: "39b5ffd7-804d-47a4-bc5b-f000f7a425a5"). InnerVolumeSpecName "kube-api-access-shnwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.116175 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.116240 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.116313 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shnwm\" (UniqueName: \"kubernetes.io/projected/39b5ffd7-804d-47a4-bc5b-f000f7a425a5-kube-api-access-shnwm\") on node \"crc\" DevicePath \"\"" Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.116430 4750 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.116518 4750 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.117246 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:07.116477826 +0000 UTC m=+1128.458950224 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "webhook-server-cert" not found Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.117300 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:07.117270018 +0000 UTC m=+1128.459742416 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "metrics-server-cert" not found Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.458123 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551358-gqjcz"] Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.468400 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551358-gqjcz"] Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.586412 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" event={"ID":"a91098d3-ee59-4f84-ad48-ee7613a8c7da","Type":"ContainerStarted","Data":"17cd4cb1e8141d0ddc272304c62df3271871d501eaf97d67a2eec679c47fef51"} Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.589307 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" podUID="a91098d3-ee59-4f84-ad48-ee7613a8c7da" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.590741 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" event={"ID":"8c687b21-5077-4d33-af70-07b01956d904","Type":"ContainerStarted","Data":"d05340dd53dfeabb28632ab7f89daf2587a13470985c4a72881c83ce66af0d7b"} Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.593960 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968\\\"\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" podUID="8c687b21-5077-4d33-af70-07b01956d904" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.594080 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" event={"ID":"93ac04dd-4b80-4e20-9113-eadef2054d0e","Type":"ContainerStarted","Data":"238a0f1fc915760ded8cd11ce87d0ada51c067df8e90278ba385274c47141e33"} Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.595600 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:bb939885bd04593ad03af901adb77ee2a2d18529b328c23288c7cc7a2ba5282e\\\"\"" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" podUID="93ac04dd-4b80-4e20-9113-eadef2054d0e" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.603754 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" event={"ID":"6bd1943b-5152-4d4e-9fa9-8c2f8ea42072","Type":"ContainerStarted","Data":"a3775aa1917dc4cfbb9795a342a67ae57ac0a29a52fde1dedeede0cd3c454815"} Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.612182 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.612541 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551364-gv2jc" event={"ID":"39b5ffd7-804d-47a4-bc5b-f000f7a425a5","Type":"ContainerDied","Data":"cbb78a283e8345ae350fd9a903e14580bbe6d688a5a42f8b1a3e4892eaa76b63"} Mar 09 18:44:05 crc kubenswrapper[4750]: I0309 18:44:05.612567 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbb78a283e8345ae350fd9a903e14580bbe6d688a5a42f8b1a3e4892eaa76b63" Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.613891 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:2d59045b8d8e6f9c5483c4fdda7c5057218d553200dc4bcf26789980ac1d9abd\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" podUID="c42f78da-7508-4f90-bee6-c05e5cbd0f86" Mar 09 18:44:05 crc kubenswrapper[4750]: E0309 18:44:05.613958 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:9f73c84a9581b5739d8da333c7b64403d7b7ca284b22c624d0effe07f3d2819c\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" podUID="1507d1a6-11a2-4a35-9c3c-1c360e4c0b32" Mar 09 18:44:06 crc kubenswrapper[4750]: I0309 18:44:06.237870 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:06 crc kubenswrapper[4750]: E0309 18:44:06.238096 4750 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:06 crc kubenswrapper[4750]: E0309 18:44:06.238188 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert podName:6f320add-8d8a-49c6-a8bd-c286772ee907 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:10.238167975 +0000 UTC m=+1131.580640373 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert") pod "infra-operator-controller-manager-f7fcc58b9-8pvfn" (UID: "6f320add-8d8a-49c6-a8bd-c286772ee907") : secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:06 crc kubenswrapper[4750]: E0309 18:44:06.662363 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:bb939885bd04593ad03af901adb77ee2a2d18529b328c23288c7cc7a2ba5282e\\\"\"" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" podUID="93ac04dd-4b80-4e20-9113-eadef2054d0e" Mar 09 18:44:06 crc kubenswrapper[4750]: E0309 18:44:06.662975 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968\\\"\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" podUID="8c687b21-5077-4d33-af70-07b01956d904" Mar 09 18:44:06 crc kubenswrapper[4750]: E0309 18:44:06.663058 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" podUID="a91098d3-ee59-4f84-ad48-ee7613a8c7da" Mar 09 18:44:06 crc kubenswrapper[4750]: I0309 18:44:06.769302 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:06 crc kubenswrapper[4750]: E0309 18:44:06.770263 4750 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:06 crc kubenswrapper[4750]: E0309 18:44:06.770317 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert podName:89799193-9df0-4505-8116-d49a86d33add nodeName:}" failed. No retries permitted until 2026-03-09 18:44:10.77030369 +0000 UTC m=+1132.112776088 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" (UID: "89799193-9df0-4505-8116-d49a86d33add") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:07 crc kubenswrapper[4750]: I0309 18:44:07.175121 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:07 crc kubenswrapper[4750]: I0309 18:44:07.175215 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:07 crc kubenswrapper[4750]: E0309 18:44:07.175368 4750 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 09 18:44:07 crc kubenswrapper[4750]: E0309 18:44:07.175423 4750 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 09 18:44:07 crc kubenswrapper[4750]: E0309 18:44:07.175496 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:11.175474892 +0000 UTC m=+1132.517947300 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "metrics-server-cert" not found Mar 09 18:44:07 crc kubenswrapper[4750]: E0309 18:44:07.175524 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:11.175513123 +0000 UTC m=+1132.517985531 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "webhook-server-cert" not found Mar 09 18:44:07 crc kubenswrapper[4750]: I0309 18:44:07.385774 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9c03a3-e6ea-411a-a9b7-99a98c41492f" path="/var/lib/kubelet/pods/8f9c03a3-e6ea-411a-a9b7-99a98c41492f/volumes" Mar 09 18:44:10 crc kubenswrapper[4750]: I0309 18:44:10.338900 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:10 crc kubenswrapper[4750]: E0309 18:44:10.339138 4750 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:10 crc kubenswrapper[4750]: E0309 18:44:10.339247 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert podName:6f320add-8d8a-49c6-a8bd-c286772ee907 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:18.339222153 +0000 UTC m=+1139.681694561 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert") pod "infra-operator-controller-manager-f7fcc58b9-8pvfn" (UID: "6f320add-8d8a-49c6-a8bd-c286772ee907") : secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:10 crc kubenswrapper[4750]: I0309 18:44:10.847013 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:10 crc kubenswrapper[4750]: E0309 18:44:10.847303 4750 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:10 crc kubenswrapper[4750]: E0309 18:44:10.847413 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert podName:89799193-9df0-4505-8116-d49a86d33add nodeName:}" failed. No retries permitted until 2026-03-09 18:44:18.847386536 +0000 UTC m=+1140.189859004 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" (UID: "89799193-9df0-4505-8116-d49a86d33add") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:11 crc kubenswrapper[4750]: I0309 18:44:11.252781 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:11 crc kubenswrapper[4750]: I0309 18:44:11.252873 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:11 crc kubenswrapper[4750]: E0309 18:44:11.253045 4750 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 09 18:44:11 crc kubenswrapper[4750]: E0309 18:44:11.253145 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:19.253123832 +0000 UTC m=+1140.595596230 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "metrics-server-cert" not found Mar 09 18:44:11 crc kubenswrapper[4750]: E0309 18:44:11.253164 4750 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 09 18:44:11 crc kubenswrapper[4750]: E0309 18:44:11.253255 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:19.253237245 +0000 UTC m=+1140.595709643 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "webhook-server-cert" not found Mar 09 18:44:17 crc kubenswrapper[4750]: E0309 18:44:17.630168 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/openstack-k8s-operators/watcher-operator:2e035aad6e396aeb72cc6aec8684c43e59f8b674" Mar 09 18:44:17 crc kubenswrapper[4750]: E0309 18:44:17.630236 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/openstack-k8s-operators/watcher-operator:2e035aad6e396aeb72cc6aec8684c43e59f8b674" Mar 09 18:44:17 crc kubenswrapper[4750]: E0309 18:44:17.631058 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.110:5001/openstack-k8s-operators/watcher-operator:2e035aad6e396aeb72cc6aec8684c43e59f8b674,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hzmpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-668c5c65dc-rt6bl_openstack-operators(6bd1943b-5152-4d4e-9fa9-8c2f8ea42072): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:44:17 crc kubenswrapper[4750]: E0309 18:44:17.632839 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" podUID="6bd1943b-5152-4d4e-9fa9-8c2f8ea42072" Mar 09 18:44:17 crc kubenswrapper[4750]: E0309 18:44:17.756141 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/openstack-k8s-operators/watcher-operator:2e035aad6e396aeb72cc6aec8684c43e59f8b674\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" podUID="6bd1943b-5152-4d4e-9fa9-8c2f8ea42072" Mar 09 18:44:18 crc kubenswrapper[4750]: I0309 18:44:18.380133 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.380809 4750 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.380895 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert podName:6f320add-8d8a-49c6-a8bd-c286772ee907 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:34.380870488 +0000 UTC m=+1155.723342886 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert") pod "infra-operator-controller-manager-f7fcc58b9-8pvfn" (UID: "6f320add-8d8a-49c6-a8bd-c286772ee907") : secret "infra-operator-webhook-server-cert" not found Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.404880 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:172f24bd4603ac3498536a8a2c8fffb07cf9113dd52bc132778ea0aa275c6b84" Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.405166 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:172f24bd4603ac3498536a8a2c8fffb07cf9113dd52bc132778ea0aa275c6b84,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fsrq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-74b6b5dc96-qp844_openstack-operators(14f86f92-1676-44e4-a56c-e611b59afe30): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.406482 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" podUID="14f86f92-1676-44e4-a56c-e611b59afe30" Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.741830 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:172f24bd4603ac3498536a8a2c8fffb07cf9113dd52bc132778ea0aa275c6b84\\\"\"" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" podUID="14f86f92-1676-44e4-a56c-e611b59afe30" Mar 09 18:44:18 crc kubenswrapper[4750]: I0309 18:44:18.890756 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.891237 4750 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.891287 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert podName:89799193-9df0-4505-8116-d49a86d33add nodeName:}" failed. No retries permitted until 2026-03-09 18:44:34.891269259 +0000 UTC m=+1156.233741657 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" (UID: "89799193-9df0-4505-8116-d49a86d33add") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.922964 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:9d723ab33964ee44704eed3223b64e828349d45dee04695434a6fcf4b6807d4c" Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.923199 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:9d723ab33964ee44704eed3223b64e828349d45dee04695434a6fcf4b6807d4c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47l5v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7c789f89c6-j799f_openstack-operators(a4506343-59e4-4442-8894-fdb158c82316): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:44:18 crc kubenswrapper[4750]: E0309 18:44:18.924452 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" podUID="a4506343-59e4-4442-8894-fdb158c82316" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.298974 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.299100 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:19 crc kubenswrapper[4750]: E0309 18:44:19.299326 4750 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 09 18:44:19 crc kubenswrapper[4750]: E0309 18:44:19.299413 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs podName:874a050d-1e92-4087-a655-b6fdb66a47e9 nodeName:}" failed. No retries permitted until 2026-03-09 18:44:35.29937574 +0000 UTC m=+1156.641848138 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs") pod "openstack-operator-controller-manager-64797568c9-p55cp" (UID: "874a050d-1e92-4087-a655-b6fdb66a47e9") : secret "metrics-server-cert" not found Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.316356 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-webhook-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.756255 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" event={"ID":"f4e026ba-f7dd-4c14-b09e-e00e6cfae20d","Type":"ContainerStarted","Data":"7606934d4567381d5dff57768fdecedaa5c0ffca8fd44d14fed703bcf3f0f5d0"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.756372 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.761991 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" event={"ID":"d8208475-d487-47ee-8f03-fa4379d3c917","Type":"ContainerStarted","Data":"b1301743f018a42365d73cbe4c05b54a9609867e7d3dd0eb44c41d8a7c8d0694"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.762129 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.767531 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" event={"ID":"3f316fad-80d2-46fb-ab97-d3f2b0559387","Type":"ContainerStarted","Data":"9c72093dae5fd276d90a070b59c78fc0c0a6eea7e4390ee9f12279c9c5780300"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.767694 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.772083 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" event={"ID":"1dc593d7-29a0-4a18-aecc-81f22e19e9b9","Type":"ContainerStarted","Data":"63a630405983fcab4a7f840b45ca2f47c31a0ba923f202b0e08120e60355637e"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.772210 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.799864 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" podStartSLOduration=2.897721272 podStartE2EDuration="17.799832195s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:03.987388669 +0000 UTC m=+1125.329861067" lastFinishedPulling="2026-03-09 18:44:18.889499582 +0000 UTC m=+1140.231971990" observedRunningTime="2026-03-09 18:44:19.796840305 +0000 UTC m=+1141.139312703" watchObservedRunningTime="2026-03-09 18:44:19.799832195 +0000 UTC m=+1141.142304603" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.804325 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" event={"ID":"434cb9b7-d90c-4081-8232-004bbd4cb8b2","Type":"ContainerStarted","Data":"95b2f25a11278983d21b88847fb221477286eeda09a030682aa0d22f476f4d6b"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.805552 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.840025 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" event={"ID":"607af51a-435c-4f0a-99c2-12c3cf6b2213","Type":"ContainerStarted","Data":"76c8eedced2760711698d0e774056acb0a062d12c0098b09dd1283d2a6435355"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.841033 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.850896 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" podStartSLOduration=2.949941831 podStartE2EDuration="17.84700273s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:03.992344221 +0000 UTC m=+1125.334816609" lastFinishedPulling="2026-03-09 18:44:18.88940511 +0000 UTC m=+1140.231877508" observedRunningTime="2026-03-09 18:44:19.840172087 +0000 UTC m=+1141.182644485" watchObservedRunningTime="2026-03-09 18:44:19.84700273 +0000 UTC m=+1141.189475128" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.872374 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" event={"ID":"e403dea7-17d0-40db-9b29-e50f33175503","Type":"ContainerStarted","Data":"475427007f07686532e916e9cb318b7c8794eeaa12f1c5c19306796c778c7415"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.872552 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.912117 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" podStartSLOduration=3.173438444 podStartE2EDuration="17.912095945s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.152607308 +0000 UTC m=+1125.495079706" lastFinishedPulling="2026-03-09 18:44:18.891264809 +0000 UTC m=+1140.233737207" observedRunningTime="2026-03-09 18:44:19.887622009 +0000 UTC m=+1141.230094407" watchObservedRunningTime="2026-03-09 18:44:19.912095945 +0000 UTC m=+1141.254568343" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.913238 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" event={"ID":"c7f102b3-0989-4582-91d1-49a22c261371","Type":"ContainerStarted","Data":"e3a088891ae24b7c551ad6ca56d8c09b6cf76ac856a4eca375e89b616701a1a1"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.914268 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.926807 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" podStartSLOduration=3.180271847 podStartE2EDuration="17.926766449s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.141472099 +0000 UTC m=+1125.483944497" lastFinishedPulling="2026-03-09 18:44:18.887966701 +0000 UTC m=+1140.230439099" observedRunningTime="2026-03-09 18:44:19.913321158 +0000 UTC m=+1141.255793556" watchObservedRunningTime="2026-03-09 18:44:19.926766449 +0000 UTC m=+1141.269238847" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.930125 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" event={"ID":"ccf58e0b-b6b3-443a-b215-087bb55da05c","Type":"ContainerStarted","Data":"44951e338ff176d44c070be1f18b5969fec10550f22daf08685f429261ebf065"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.930940 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.939369 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" podStartSLOduration=3.038454734 podStartE2EDuration="17.939354596s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:03.48383807 +0000 UTC m=+1124.826310468" lastFinishedPulling="2026-03-09 18:44:18.384737932 +0000 UTC m=+1139.727210330" observedRunningTime="2026-03-09 18:44:19.935121563 +0000 UTC m=+1141.277593951" watchObservedRunningTime="2026-03-09 18:44:19.939354596 +0000 UTC m=+1141.281826994" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.954863 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" event={"ID":"bf342fef-8819-4328-9551-13d21041bff6","Type":"ContainerStarted","Data":"137eca02a17a1fa6bf4c5abe12f32840a456499c2b07b390c744b452b5c8b637"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.955662 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.992862 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" event={"ID":"9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9","Type":"ContainerStarted","Data":"596efd0117c0f41aa98d43a385c439fa14e49c017fa7cff30b3feead2101c0b2"} Mar 09 18:44:19 crc kubenswrapper[4750]: I0309 18:44:19.994020 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.015197 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" podStartSLOduration=3.034726944 podStartE2EDuration="18.015177718s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:03.90911189 +0000 UTC m=+1125.251584288" lastFinishedPulling="2026-03-09 18:44:18.889562664 +0000 UTC m=+1140.232035062" observedRunningTime="2026-03-09 18:44:20.013060971 +0000 UTC m=+1141.355533369" watchObservedRunningTime="2026-03-09 18:44:20.015177718 +0000 UTC m=+1141.357650116" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.015538 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" podStartSLOduration=3.498794915 podStartE2EDuration="18.015531978s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.375725559 +0000 UTC m=+1125.718197957" lastFinishedPulling="2026-03-09 18:44:18.892462622 +0000 UTC m=+1140.234935020" observedRunningTime="2026-03-09 18:44:19.981187158 +0000 UTC m=+1141.323659556" watchObservedRunningTime="2026-03-09 18:44:20.015531978 +0000 UTC m=+1141.358004376" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.018330 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" event={"ID":"e671bb3b-ca69-4229-9e9b-e823bfcf2f92","Type":"ContainerStarted","Data":"766f3d63cf09cede88baf6f4a3da38facb2ecf0fdf575e21f3e7551ecbf8189d"} Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.018462 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" Mar 09 18:44:20 crc kubenswrapper[4750]: E0309 18:44:20.025307 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:9d723ab33964ee44704eed3223b64e828349d45dee04695434a6fcf4b6807d4c\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" podUID="a4506343-59e4-4442-8894-fdb158c82316" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.062953 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" podStartSLOduration=2.7948586239999997 podStartE2EDuration="18.062926668s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:03.621715426 +0000 UTC m=+1124.964187824" lastFinishedPulling="2026-03-09 18:44:18.88978347 +0000 UTC m=+1140.232255868" observedRunningTime="2026-03-09 18:44:20.048044219 +0000 UTC m=+1141.390516617" watchObservedRunningTime="2026-03-09 18:44:20.062926668 +0000 UTC m=+1141.405399066" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.083658 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" podStartSLOduration=2.83570988 podStartE2EDuration="18.083636954s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:03.642660438 +0000 UTC m=+1124.985132836" lastFinishedPulling="2026-03-09 18:44:18.890587512 +0000 UTC m=+1140.233059910" observedRunningTime="2026-03-09 18:44:20.081948479 +0000 UTC m=+1141.424420877" watchObservedRunningTime="2026-03-09 18:44:20.083636954 +0000 UTC m=+1141.426109352" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.115312 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" podStartSLOduration=3.195173006 podStartE2EDuration="18.115291913s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:03.969322004 +0000 UTC m=+1125.311794402" lastFinishedPulling="2026-03-09 18:44:18.889440911 +0000 UTC m=+1140.231913309" observedRunningTime="2026-03-09 18:44:20.112466806 +0000 UTC m=+1141.454939214" watchObservedRunningTime="2026-03-09 18:44:20.115291913 +0000 UTC m=+1141.457764311" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.135419 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" podStartSLOduration=3.62096588 podStartE2EDuration="18.135402801s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.375293027 +0000 UTC m=+1125.717765425" lastFinishedPulling="2026-03-09 18:44:18.889729948 +0000 UTC m=+1140.232202346" observedRunningTime="2026-03-09 18:44:20.133456699 +0000 UTC m=+1141.475929097" watchObservedRunningTime="2026-03-09 18:44:20.135402801 +0000 UTC m=+1141.477875199" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.159921 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" podStartSLOduration=3.417961698 podStartE2EDuration="18.159896718s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.148390645 +0000 UTC m=+1125.490863043" lastFinishedPulling="2026-03-09 18:44:18.890325665 +0000 UTC m=+1140.232798063" observedRunningTime="2026-03-09 18:44:20.155145341 +0000 UTC m=+1141.497617749" watchObservedRunningTime="2026-03-09 18:44:20.159896718 +0000 UTC m=+1141.502369116" Mar 09 18:44:20 crc kubenswrapper[4750]: I0309 18:44:20.695530 4750 scope.go:117] "RemoveContainer" containerID="e7928fd87977ab59a664baed25da0b8c02b15737bceb8fc18cace73eddd3f19a" Mar 09 18:44:27 crc kubenswrapper[4750]: I0309 18:44:27.089837 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" event={"ID":"93ac04dd-4b80-4e20-9113-eadef2054d0e","Type":"ContainerStarted","Data":"c90a9cb1a83dcfda8bbc05cdd233e6a8ec34775fee841e2b16311a1841962c1d"} Mar 09 18:44:27 crc kubenswrapper[4750]: I0309 18:44:27.091468 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" Mar 09 18:44:27 crc kubenswrapper[4750]: I0309 18:44:27.111997 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" podStartSLOduration=3.245265359 podStartE2EDuration="25.111975244s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.595711526 +0000 UTC m=+1125.938183934" lastFinishedPulling="2026-03-09 18:44:26.462421421 +0000 UTC m=+1147.804893819" observedRunningTime="2026-03-09 18:44:27.105177112 +0000 UTC m=+1148.447649520" watchObservedRunningTime="2026-03-09 18:44:27.111975244 +0000 UTC m=+1148.454447642" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.130365 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" event={"ID":"8c687b21-5077-4d33-af70-07b01956d904","Type":"ContainerStarted","Data":"78646cb3df6b996894fc447cc23ecd9d5fc9b414b6d063f14b98c0b33cd485a2"} Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.130900 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.132649 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" event={"ID":"6bd1943b-5152-4d4e-9fa9-8c2f8ea42072","Type":"ContainerStarted","Data":"8d3d61e66750620bbafec0944759b766d2b3ca02a0bcc366dedcb254515cf04b"} Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.132888 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.134964 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" event={"ID":"1507d1a6-11a2-4a35-9c3c-1c360e4c0b32","Type":"ContainerStarted","Data":"f18c83fa710cf54d6ab7899cc64792242bf4b28fad572d077188413d64fe5e63"} Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.135164 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.136701 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" event={"ID":"c42f78da-7508-4f90-bee6-c05e5cbd0f86","Type":"ContainerStarted","Data":"225b4dcb8fbfe1f6c3d1b6cabd88d7a455d2987c28d2624dcafd362230e35187"} Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.136908 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.138836 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" event={"ID":"a91098d3-ee59-4f84-ad48-ee7613a8c7da","Type":"ContainerStarted","Data":"4cff42adca861b4ea8482fe02ffcdf9ba10d9dd16d12c5c9db2b065d7e6a8ecc"} Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.170114 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" podStartSLOduration=3.79591002 podStartE2EDuration="29.170083s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.601787059 +0000 UTC m=+1125.944259457" lastFinishedPulling="2026-03-09 18:44:29.975960039 +0000 UTC m=+1151.318432437" observedRunningTime="2026-03-09 18:44:31.155774087 +0000 UTC m=+1152.498246495" watchObservedRunningTime="2026-03-09 18:44:31.170083 +0000 UTC m=+1152.512555408" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.189667 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" podStartSLOduration=3.727856646 podStartE2EDuration="29.189644775s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.59511413 +0000 UTC m=+1125.937586518" lastFinishedPulling="2026-03-09 18:44:30.056902249 +0000 UTC m=+1151.399374647" observedRunningTime="2026-03-09 18:44:31.182904224 +0000 UTC m=+1152.525376652" watchObservedRunningTime="2026-03-09 18:44:31.189644775 +0000 UTC m=+1152.532117183" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.211277 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fbjb2" podStartSLOduration=2.764611571 podStartE2EDuration="28.211257094s" podCreationTimestamp="2026-03-09 18:44:03 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.606043993 +0000 UTC m=+1125.948516391" lastFinishedPulling="2026-03-09 18:44:30.052689516 +0000 UTC m=+1151.395161914" observedRunningTime="2026-03-09 18:44:31.206430795 +0000 UTC m=+1152.548903203" watchObservedRunningTime="2026-03-09 18:44:31.211257094 +0000 UTC m=+1152.553729502" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.227621 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" podStartSLOduration=3.638902851 podStartE2EDuration="29.227598882s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.386544959 +0000 UTC m=+1125.729017357" lastFinishedPulling="2026-03-09 18:44:29.975241 +0000 UTC m=+1151.317713388" observedRunningTime="2026-03-09 18:44:31.221897389 +0000 UTC m=+1152.564369797" watchObservedRunningTime="2026-03-09 18:44:31.227598882 +0000 UTC m=+1152.570071290" Mar 09 18:44:31 crc kubenswrapper[4750]: I0309 18:44:31.245785 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" podStartSLOduration=3.584493913 podStartE2EDuration="29.245761039s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.386546539 +0000 UTC m=+1125.729018937" lastFinishedPulling="2026-03-09 18:44:30.047813665 +0000 UTC m=+1151.390286063" observedRunningTime="2026-03-09 18:44:31.242901462 +0000 UTC m=+1152.585373870" watchObservedRunningTime="2026-03-09 18:44:31.245761039 +0000 UTC m=+1152.588233447" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.568199 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-qdwbn" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.578102 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-86r5x" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.641250 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-czkhf" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.654913 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-xrjb7" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.698443 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-wl9k2" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.774143 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-tkcr6" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.855318 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-lmj58" Mar 09 18:44:32 crc kubenswrapper[4750]: I0309 18:44:32.940120 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-67d996989d-wrkj4" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.010119 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7b6bfb6475-znrqs" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.010900 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54688575f-n8ldk" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.154337 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" event={"ID":"14f86f92-1676-44e4-a56c-e611b59afe30","Type":"ContainerStarted","Data":"b60d2bf9a6c05b96335a4adc4709488b893db4decdb7876b288d3f483edac91c"} Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.154769 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.156150 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" event={"ID":"a4506343-59e4-4442-8894-fdb158c82316","Type":"ContainerStarted","Data":"437038aa7d9e7a76db163aa77dfd1caee18e26b5e88d3fd65486491479318554"} Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.156291 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.175383 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" podStartSLOduration=2.60967489 podStartE2EDuration="31.175361166s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.209555334 +0000 UTC m=+1125.552027732" lastFinishedPulling="2026-03-09 18:44:32.77524161 +0000 UTC m=+1154.117714008" observedRunningTime="2026-03-09 18:44:33.168484161 +0000 UTC m=+1154.510956559" watchObservedRunningTime="2026-03-09 18:44:33.175361166 +0000 UTC m=+1154.517833574" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.191707 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" podStartSLOduration=2.474100387 podStartE2EDuration="31.191679454s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:04.151123888 +0000 UTC m=+1125.493596286" lastFinishedPulling="2026-03-09 18:44:32.868702955 +0000 UTC m=+1154.211175353" observedRunningTime="2026-03-09 18:44:33.186474544 +0000 UTC m=+1154.528946932" watchObservedRunningTime="2026-03-09 18:44:33.191679454 +0000 UTC m=+1154.534151852" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.264348 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-2l9js" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.341417 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-skdbr" Mar 09 18:44:33 crc kubenswrapper[4750]: I0309 18:44:33.391335 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-rvpms" Mar 09 18:44:34 crc kubenswrapper[4750]: I0309 18:44:34.453416 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:34 crc kubenswrapper[4750]: I0309 18:44:34.471608 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f320add-8d8a-49c6-a8bd-c286772ee907-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-8pvfn\" (UID: \"6f320add-8d8a-49c6-a8bd-c286772ee907\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:34 crc kubenswrapper[4750]: I0309 18:44:34.637960 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gpjhk" Mar 09 18:44:34 crc kubenswrapper[4750]: I0309 18:44:34.646857 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:34 crc kubenswrapper[4750]: I0309 18:44:34.967293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:34 crc kubenswrapper[4750]: I0309 18:44:34.977335 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/89799193-9df0-4505-8116-d49a86d33add-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw\" (UID: \"89799193-9df0-4505-8116-d49a86d33add\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:34 crc kubenswrapper[4750]: W0309 18:44:34.998881 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f320add_8d8a_49c6_a8bd_c286772ee907.slice/crio-c30face44da93934d198bb9549cb3912ac3eb7a5138b9d3212795770bab46568 WatchSource:0}: Error finding container c30face44da93934d198bb9549cb3912ac3eb7a5138b9d3212795770bab46568: Status 404 returned error can't find the container with id c30face44da93934d198bb9549cb3912ac3eb7a5138b9d3212795770bab46568 Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.003021 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn"] Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.075094 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hmj8h" Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.083685 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.179075 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" event={"ID":"6f320add-8d8a-49c6-a8bd-c286772ee907","Type":"ContainerStarted","Data":"c30face44da93934d198bb9549cb3912ac3eb7a5138b9d3212795770bab46568"} Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.321727 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw"] Mar 09 18:44:35 crc kubenswrapper[4750]: W0309 18:44:35.326221 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89799193_9df0_4505_8116_d49a86d33add.slice/crio-8c1ed0698b5d60913a7a0320040a8382ae864af677595741ca726648237ddbb1 WatchSource:0}: Error finding container 8c1ed0698b5d60913a7a0320040a8382ae864af677595741ca726648237ddbb1: Status 404 returned error can't find the container with id 8c1ed0698b5d60913a7a0320040a8382ae864af677595741ca726648237ddbb1 Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.373441 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.378330 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/874a050d-1e92-4087-a655-b6fdb66a47e9-metrics-certs\") pod \"openstack-operator-controller-manager-64797568c9-p55cp\" (UID: \"874a050d-1e92-4087-a655-b6fdb66a47e9\") " pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.453092 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zc9z6" Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.461773 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:35 crc kubenswrapper[4750]: I0309 18:44:35.732371 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp"] Mar 09 18:44:36 crc kubenswrapper[4750]: I0309 18:44:36.187957 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" event={"ID":"89799193-9df0-4505-8116-d49a86d33add","Type":"ContainerStarted","Data":"8c1ed0698b5d60913a7a0320040a8382ae864af677595741ca726648237ddbb1"} Mar 09 18:44:36 crc kubenswrapper[4750]: I0309 18:44:36.190383 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" event={"ID":"874a050d-1e92-4087-a655-b6fdb66a47e9","Type":"ContainerStarted","Data":"7934902f7527873ef6e91be9f18eae2ff2db53543fd94bf9b92d8bcc4455b0be"} Mar 09 18:44:36 crc kubenswrapper[4750]: I0309 18:44:36.190410 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" event={"ID":"874a050d-1e92-4087-a655-b6fdb66a47e9","Type":"ContainerStarted","Data":"4bf783a7a96092e8c7f80e1b16f7fa23e72f5ead84695d31ce65150a7aa5df4a"} Mar 09 18:44:36 crc kubenswrapper[4750]: I0309 18:44:36.191521 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:44:36 crc kubenswrapper[4750]: I0309 18:44:36.227620 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" podStartSLOduration=34.227594298 podStartE2EDuration="34.227594298s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:44:36.218127904 +0000 UTC m=+1157.560600302" watchObservedRunningTime="2026-03-09 18:44:36.227594298 +0000 UTC m=+1157.570066696" Mar 09 18:44:38 crc kubenswrapper[4750]: I0309 18:44:38.209789 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" event={"ID":"89799193-9df0-4505-8116-d49a86d33add","Type":"ContainerStarted","Data":"582f7ce3ab4cc7ff8f6e94d54b8804c5fa7a0eb5ce0922c322b556c73ccb4103"} Mar 09 18:44:38 crc kubenswrapper[4750]: I0309 18:44:38.210819 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:38 crc kubenswrapper[4750]: I0309 18:44:38.212856 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" event={"ID":"6f320add-8d8a-49c6-a8bd-c286772ee907","Type":"ContainerStarted","Data":"af8e44ee344c528f79b4ae5512cc3dbf39813652af7c999a143cd1453e067f00"} Mar 09 18:44:38 crc kubenswrapper[4750]: I0309 18:44:38.248716 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" podStartSLOduration=34.022201202 podStartE2EDuration="36.248690378s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:35.328829374 +0000 UTC m=+1156.671301772" lastFinishedPulling="2026-03-09 18:44:37.55531855 +0000 UTC m=+1158.897790948" observedRunningTime="2026-03-09 18:44:38.24505494 +0000 UTC m=+1159.587527378" watchObservedRunningTime="2026-03-09 18:44:38.248690378 +0000 UTC m=+1159.591162806" Mar 09 18:44:38 crc kubenswrapper[4750]: I0309 18:44:38.288487 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" podStartSLOduration=33.7401322 podStartE2EDuration="36.288442523s" podCreationTimestamp="2026-03-09 18:44:02 +0000 UTC" firstStartedPulling="2026-03-09 18:44:35.001690535 +0000 UTC m=+1156.344162933" lastFinishedPulling="2026-03-09 18:44:37.550000858 +0000 UTC m=+1158.892473256" observedRunningTime="2026-03-09 18:44:38.286257354 +0000 UTC m=+1159.628729772" watchObservedRunningTime="2026-03-09 18:44:38.288442523 +0000 UTC m=+1159.630914931" Mar 09 18:44:39 crc kubenswrapper[4750]: I0309 18:44:39.221124 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:42 crc kubenswrapper[4750]: I0309 18:44:42.896233 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7c789f89c6-j799f" Mar 09 18:44:43 crc kubenswrapper[4750]: I0309 18:44:43.151160 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-qp844" Mar 09 18:44:43 crc kubenswrapper[4750]: I0309 18:44:43.164620 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-dnt7g" Mar 09 18:44:43 crc kubenswrapper[4750]: I0309 18:44:43.206062 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-hj5hw" Mar 09 18:44:43 crc kubenswrapper[4750]: I0309 18:44:43.536252 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-z96st" Mar 09 18:44:43 crc kubenswrapper[4750]: I0309 18:44:43.560019 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-668c5c65dc-rt6bl" Mar 09 18:44:44 crc kubenswrapper[4750]: I0309 18:44:44.655593 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-8pvfn" Mar 09 18:44:45 crc kubenswrapper[4750]: I0309 18:44:45.092566 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw" Mar 09 18:44:45 crc kubenswrapper[4750]: I0309 18:44:45.471723 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-64797568c9-p55cp" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.166814 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6"] Mar 09 18:45:00 crc kubenswrapper[4750]: E0309 18:45:00.168168 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39b5ffd7-804d-47a4-bc5b-f000f7a425a5" containerName="oc" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.168185 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="39b5ffd7-804d-47a4-bc5b-f000f7a425a5" containerName="oc" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.168368 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="39b5ffd7-804d-47a4-bc5b-f000f7a425a5" containerName="oc" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.169095 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.176838 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.177040 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.185681 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6"] Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.223792 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4jld\" (UniqueName: \"kubernetes.io/projected/53eb215e-071b-49e8-b1c7-a97471da3fa2-kube-api-access-w4jld\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.223997 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53eb215e-071b-49e8-b1c7-a97471da3fa2-config-volume\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.224064 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53eb215e-071b-49e8-b1c7-a97471da3fa2-secret-volume\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.325136 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4jld\" (UniqueName: \"kubernetes.io/projected/53eb215e-071b-49e8-b1c7-a97471da3fa2-kube-api-access-w4jld\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.325232 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53eb215e-071b-49e8-b1c7-a97471da3fa2-config-volume\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.325267 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53eb215e-071b-49e8-b1c7-a97471da3fa2-secret-volume\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.326367 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53eb215e-071b-49e8-b1c7-a97471da3fa2-config-volume\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.332396 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53eb215e-071b-49e8-b1c7-a97471da3fa2-secret-volume\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.345950 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4jld\" (UniqueName: \"kubernetes.io/projected/53eb215e-071b-49e8-b1c7-a97471da3fa2-kube-api-access-w4jld\") pod \"collect-profiles-29551365-xklq6\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.492862 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:00 crc kubenswrapper[4750]: I0309 18:45:00.742760 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6"] Mar 09 18:45:01 crc kubenswrapper[4750]: I0309 18:45:01.440098 4750 generic.go:334] "Generic (PLEG): container finished" podID="53eb215e-071b-49e8-b1c7-a97471da3fa2" containerID="9d6bbc169c6d6e063ea2e716563a12ac31e552479910608141017e9225fa94ab" exitCode=0 Mar 09 18:45:01 crc kubenswrapper[4750]: I0309 18:45:01.440168 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" event={"ID":"53eb215e-071b-49e8-b1c7-a97471da3fa2","Type":"ContainerDied","Data":"9d6bbc169c6d6e063ea2e716563a12ac31e552479910608141017e9225fa94ab"} Mar 09 18:45:01 crc kubenswrapper[4750]: I0309 18:45:01.440593 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" event={"ID":"53eb215e-071b-49e8-b1c7-a97471da3fa2","Type":"ContainerStarted","Data":"1c63a016d60547351e129fd60e5bf36fad50778fd91ddf9834df6eb0844ece19"} Mar 09 18:45:02 crc kubenswrapper[4750]: I0309 18:45:02.830309 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:02 crc kubenswrapper[4750]: I0309 18:45:02.977441 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53eb215e-071b-49e8-b1c7-a97471da3fa2-config-volume\") pod \"53eb215e-071b-49e8-b1c7-a97471da3fa2\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " Mar 09 18:45:02 crc kubenswrapper[4750]: I0309 18:45:02.977522 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53eb215e-071b-49e8-b1c7-a97471da3fa2-secret-volume\") pod \"53eb215e-071b-49e8-b1c7-a97471da3fa2\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " Mar 09 18:45:02 crc kubenswrapper[4750]: I0309 18:45:02.977588 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4jld\" (UniqueName: \"kubernetes.io/projected/53eb215e-071b-49e8-b1c7-a97471da3fa2-kube-api-access-w4jld\") pod \"53eb215e-071b-49e8-b1c7-a97471da3fa2\" (UID: \"53eb215e-071b-49e8-b1c7-a97471da3fa2\") " Mar 09 18:45:02 crc kubenswrapper[4750]: I0309 18:45:02.978742 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53eb215e-071b-49e8-b1c7-a97471da3fa2-config-volume" (OuterVolumeSpecName: "config-volume") pod "53eb215e-071b-49e8-b1c7-a97471da3fa2" (UID: "53eb215e-071b-49e8-b1c7-a97471da3fa2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:02 crc kubenswrapper[4750]: I0309 18:45:02.989776 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53eb215e-071b-49e8-b1c7-a97471da3fa2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "53eb215e-071b-49e8-b1c7-a97471da3fa2" (UID: "53eb215e-071b-49e8-b1c7-a97471da3fa2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:45:02 crc kubenswrapper[4750]: I0309 18:45:02.989794 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53eb215e-071b-49e8-b1c7-a97471da3fa2-kube-api-access-w4jld" (OuterVolumeSpecName: "kube-api-access-w4jld") pod "53eb215e-071b-49e8-b1c7-a97471da3fa2" (UID: "53eb215e-071b-49e8-b1c7-a97471da3fa2"). InnerVolumeSpecName "kube-api-access-w4jld". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:45:03 crc kubenswrapper[4750]: I0309 18:45:03.080588 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53eb215e-071b-49e8-b1c7-a97471da3fa2-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:03 crc kubenswrapper[4750]: I0309 18:45:03.080701 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4jld\" (UniqueName: \"kubernetes.io/projected/53eb215e-071b-49e8-b1c7-a97471da3fa2-kube-api-access-w4jld\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:03 crc kubenswrapper[4750]: I0309 18:45:03.080727 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53eb215e-071b-49e8-b1c7-a97471da3fa2-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:03 crc kubenswrapper[4750]: I0309 18:45:03.458776 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" event={"ID":"53eb215e-071b-49e8-b1c7-a97471da3fa2","Type":"ContainerDied","Data":"1c63a016d60547351e129fd60e5bf36fad50778fd91ddf9834df6eb0844ece19"} Mar 09 18:45:03 crc kubenswrapper[4750]: I0309 18:45:03.458829 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c63a016d60547351e129fd60e5bf36fad50778fd91ddf9834df6eb0844ece19" Mar 09 18:45:03 crc kubenswrapper[4750]: I0309 18:45:03.458897 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.451036 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69554c7465-vpp94"] Mar 09 18:45:05 crc kubenswrapper[4750]: E0309 18:45:05.454102 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53eb215e-071b-49e8-b1c7-a97471da3fa2" containerName="collect-profiles" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.454136 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="53eb215e-071b-49e8-b1c7-a97471da3fa2" containerName="collect-profiles" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.454345 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="53eb215e-071b-49e8-b1c7-a97471da3fa2" containerName="collect-profiles" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.455590 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.464991 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69554c7465-vpp94"] Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.467162 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.467449 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.467466 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8kgvq" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.467513 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.553140 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56d9f89cb5-6plnp"] Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.554372 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.570120 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.574486 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56d9f89cb5-6plnp"] Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.617978 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-config\") pod \"dnsmasq-dns-69554c7465-vpp94\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.618076 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhkhz\" (UniqueName: \"kubernetes.io/projected/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-kube-api-access-mhkhz\") pod \"dnsmasq-dns-69554c7465-vpp94\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.720322 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-config\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.720500 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhkhz\" (UniqueName: \"kubernetes.io/projected/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-kube-api-access-mhkhz\") pod \"dnsmasq-dns-69554c7465-vpp94\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.720564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxq46\" (UniqueName: \"kubernetes.io/projected/8cc465e5-9a2c-4774-b801-1f348194e086-kube-api-access-vxq46\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.720649 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-dns-svc\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.720685 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-config\") pod \"dnsmasq-dns-69554c7465-vpp94\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.721785 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-config\") pod \"dnsmasq-dns-69554c7465-vpp94\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.750037 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhkhz\" (UniqueName: \"kubernetes.io/projected/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-kube-api-access-mhkhz\") pod \"dnsmasq-dns-69554c7465-vpp94\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.778968 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.821814 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxq46\" (UniqueName: \"kubernetes.io/projected/8cc465e5-9a2c-4774-b801-1f348194e086-kube-api-access-vxq46\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.821908 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-dns-svc\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.821974 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-config\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.822980 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-config\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.823150 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-dns-svc\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.849183 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxq46\" (UniqueName: \"kubernetes.io/projected/8cc465e5-9a2c-4774-b801-1f348194e086-kube-api-access-vxq46\") pod \"dnsmasq-dns-56d9f89cb5-6plnp\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:05 crc kubenswrapper[4750]: I0309 18:45:05.878055 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:06 crc kubenswrapper[4750]: I0309 18:45:06.297714 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69554c7465-vpp94"] Mar 09 18:45:06 crc kubenswrapper[4750]: I0309 18:45:06.365759 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56d9f89cb5-6plnp"] Mar 09 18:45:06 crc kubenswrapper[4750]: W0309 18:45:06.372786 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cc465e5_9a2c_4774_b801_1f348194e086.slice/crio-4d3a82cfe7ec846f198b2ad120d8980a11f4fa888ac805cc7570ad9125295259 WatchSource:0}: Error finding container 4d3a82cfe7ec846f198b2ad120d8980a11f4fa888ac805cc7570ad9125295259: Status 404 returned error can't find the container with id 4d3a82cfe7ec846f198b2ad120d8980a11f4fa888ac805cc7570ad9125295259 Mar 09 18:45:06 crc kubenswrapper[4750]: I0309 18:45:06.487959 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" event={"ID":"8cc465e5-9a2c-4774-b801-1f348194e086","Type":"ContainerStarted","Data":"4d3a82cfe7ec846f198b2ad120d8980a11f4fa888ac805cc7570ad9125295259"} Mar 09 18:45:06 crc kubenswrapper[4750]: I0309 18:45:06.490069 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69554c7465-vpp94" event={"ID":"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3","Type":"ContainerStarted","Data":"a08a31fe30417068949185c62b5310be40661dd879cab24bdfe821c8ceee724c"} Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.020047 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56d9f89cb5-6plnp"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.047005 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fdf6c6f7-h2j9d"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.048518 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.062281 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fdf6c6f7-h2j9d"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.174242 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkxnw\" (UniqueName: \"kubernetes.io/projected/be777a18-ce11-4d52-852d-b7bb3ffca188-kube-api-access-bkxnw\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.174289 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-dns-svc\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.174358 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-config\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.276395 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-config\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.276510 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkxnw\" (UniqueName: \"kubernetes.io/projected/be777a18-ce11-4d52-852d-b7bb3ffca188-kube-api-access-bkxnw\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.276532 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-dns-svc\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.277413 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-dns-svc\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.277411 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-config\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.301171 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69554c7465-vpp94"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.309502 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkxnw\" (UniqueName: \"kubernetes.io/projected/be777a18-ce11-4d52-852d-b7bb3ffca188-kube-api-access-bkxnw\") pod \"dnsmasq-dns-7fdf6c6f7-h2j9d\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.333060 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b67658d95-pfl2f"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.336113 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.340526 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b67658d95-pfl2f"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.365203 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.490157 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-dns-svc\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.490229 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b9w7\" (UniqueName: \"kubernetes.io/projected/c48f9535-ad2d-4cac-a39c-043f1d25859a-kube-api-access-6b9w7\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.490252 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-config\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.574144 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fdf6c6f7-h2j9d"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.593765 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-dns-svc\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.593838 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b9w7\" (UniqueName: \"kubernetes.io/projected/c48f9535-ad2d-4cac-a39c-043f1d25859a-kube-api-access-6b9w7\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.593858 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-config\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.594715 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-config\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.595226 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-dns-svc\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.604201 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-595d94d48f-57ntm"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.609718 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.619186 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-595d94d48f-57ntm"] Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.630607 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b9w7\" (UniqueName: \"kubernetes.io/projected/c48f9535-ad2d-4cac-a39c-043f1d25859a-kube-api-access-6b9w7\") pod \"dnsmasq-dns-6b67658d95-pfl2f\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.651100 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.796671 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-dns-svc\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.797496 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68fsb\" (UniqueName: \"kubernetes.io/projected/e8ab0211-a492-45f9-9606-08706461761d-kube-api-access-68fsb\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.797559 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-config\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.898675 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-config\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.898775 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-dns-svc\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.898854 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68fsb\" (UniqueName: \"kubernetes.io/projected/e8ab0211-a492-45f9-9606-08706461761d-kube-api-access-68fsb\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.899711 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-config\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.899767 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-dns-svc\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.921866 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68fsb\" (UniqueName: \"kubernetes.io/projected/e8ab0211-a492-45f9-9606-08706461761d-kube-api-access-68fsb\") pod \"dnsmasq-dns-595d94d48f-57ntm\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:09 crc kubenswrapper[4750]: I0309 18:45:09.951414 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.185716 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.186942 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.189623 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.190101 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5prvj" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.190844 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.191077 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.191287 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.191415 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.191601 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.206619 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304599 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304717 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/378d9b58-f830-4d61-b408-26668a301507-pod-info\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304746 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304779 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304809 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-config-data\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304901 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njwwt\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-kube-api-access-njwwt\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304940 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-server-conf\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304971 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.304995 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.305056 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/378d9b58-f830-4d61-b408-26668a301507-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.305083 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406729 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/378d9b58-f830-4d61-b408-26668a301507-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406778 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406822 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406844 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406858 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/378d9b58-f830-4d61-b408-26668a301507-pod-info\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406879 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406900 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-config-data\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406935 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njwwt\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-kube-api-access-njwwt\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406958 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-server-conf\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406980 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.406996 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.407431 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.408028 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.408141 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.408258 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.408736 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-server-conf\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.408938 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-config-data\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.412288 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.413596 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/378d9b58-f830-4d61-b408-26668a301507-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.423673 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/378d9b58-f830-4d61-b408-26668a301507-pod-info\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.425025 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.432302 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njwwt\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-kube-api-access-njwwt\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.441411 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.445511 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.447858 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.451724 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.451736 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.451883 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.451742 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.452060 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.452146 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.452647 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ck6bg" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.456973 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.524375 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.614660 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617519 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617660 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617760 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3236a6fb-c288-441c-96e5-7941f818b0af-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617785 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617821 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qkgb\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-kube-api-access-6qkgb\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617889 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617922 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617936 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.617996 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3236a6fb-c288-441c-96e5-7941f818b0af-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.618267 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.726670 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3236a6fb-c288-441c-96e5-7941f818b0af-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.726748 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.726782 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qkgb\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-kube-api-access-6qkgb\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.726845 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.726884 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.726914 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.726968 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3236a6fb-c288-441c-96e5-7941f818b0af-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.727009 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.727042 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.727097 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.727165 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.727352 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.730203 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.730477 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.730495 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.730639 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.732381 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.733211 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.735090 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.735201 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3236a6fb-c288-441c-96e5-7941f818b0af-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.739155 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/notifications-rabbitmq-server-0"] Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.740665 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.743144 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"notifications-rabbitmq-erlang-cookie" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.748142 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"notifications-rabbitmq-server-dockercfg-sc2pc" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.748495 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"notifications-rabbitmq-config-data" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.748744 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"notifications-rabbitmq-server-conf" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.748853 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"notifications-rabbitmq-plugins-conf" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.748935 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"notifications-rabbitmq-default-user" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.749048 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-notifications-rabbitmq-svc" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.756075 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3236a6fb-c288-441c-96e5-7941f818b0af-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.763016 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qkgb\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-kube-api-access-6qkgb\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.765186 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/notifications-rabbitmq-server-0"] Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.782198 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.790218 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.828417 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r78lr\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-kube-api-access-r78lr\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.828837 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-config-data\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.828874 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-erlang-cookie\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.828901 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-confd\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.828944 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-server-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.829127 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3926781f-b03f-45a7-82e5-10abf0c16c1e-erlang-cookie-secret\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.829189 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-tls\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.829246 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-plugins\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.829287 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3926781f-b03f-45a7-82e5-10abf0c16c1e-pod-info\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.829320 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.829343 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-plugins-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931094 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-config-data\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931150 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-erlang-cookie\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-confd\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931208 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-server-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931231 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3926781f-b03f-45a7-82e5-10abf0c16c1e-erlang-cookie-secret\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931255 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-tls\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931291 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-plugins\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.931315 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3926781f-b03f-45a7-82e5-10abf0c16c1e-pod-info\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.932591 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.932613 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-plugins-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.932652 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r78lr\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-kube-api-access-r78lr\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.933542 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.934774 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-server-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.934803 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-config-data\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.940597 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-tls\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.941358 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3926781f-b03f-45a7-82e5-10abf0c16c1e-pod-info\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.947713 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-erlang-cookie\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.947951 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-plugins\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.948362 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3926781f-b03f-45a7-82e5-10abf0c16c1e-plugins-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.950344 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3926781f-b03f-45a7-82e5-10abf0c16c1e-erlang-cookie-secret\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.963046 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-rabbitmq-confd\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.969537 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r78lr\" (UniqueName: \"kubernetes.io/projected/3926781f-b03f-45a7-82e5-10abf0c16c1e-kube-api-access-r78lr\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:10 crc kubenswrapper[4750]: I0309 18:45:10.973244 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"3926781f-b03f-45a7-82e5-10abf0c16c1e\") " pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:11 crc kubenswrapper[4750]: I0309 18:45:11.143769 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.023329 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.025897 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.035800 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-684rt" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.036057 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.041313 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.043504 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.043980 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.044127 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.156862 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89tht\" (UniqueName: \"kubernetes.io/projected/302428aa-7eba-4f8b-8868-1e8e883d38c2-kube-api-access-89tht\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.156969 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-config-data-default\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.157177 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.157244 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/302428aa-7eba-4f8b-8868-1e8e883d38c2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.157320 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.157810 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/302428aa-7eba-4f8b-8868-1e8e883d38c2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.157935 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302428aa-7eba-4f8b-8868-1e8e883d38c2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.157994 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-kolla-config\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.259875 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89tht\" (UniqueName: \"kubernetes.io/projected/302428aa-7eba-4f8b-8868-1e8e883d38c2-kube-api-access-89tht\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.259962 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-config-data-default\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.260014 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.260056 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/302428aa-7eba-4f8b-8868-1e8e883d38c2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.260095 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.260148 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/302428aa-7eba-4f8b-8868-1e8e883d38c2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.260192 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302428aa-7eba-4f8b-8868-1e8e883d38c2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.260227 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-kolla-config\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.260919 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.261069 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/302428aa-7eba-4f8b-8868-1e8e883d38c2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.261350 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-kolla-config\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.262519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-config-data-default\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.264323 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/302428aa-7eba-4f8b-8868-1e8e883d38c2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.267213 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/302428aa-7eba-4f8b-8868-1e8e883d38c2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.267786 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302428aa-7eba-4f8b-8868-1e8e883d38c2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.280338 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89tht\" (UniqueName: \"kubernetes.io/projected/302428aa-7eba-4f8b-8868-1e8e883d38c2-kube-api-access-89tht\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.294956 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"302428aa-7eba-4f8b-8868-1e8e883d38c2\") " pod="openstack/openstack-galera-0" Mar 09 18:45:12 crc kubenswrapper[4750]: I0309 18:45:12.358314 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.511899 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.514000 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.518359 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.518605 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.518770 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-bgxvs" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.522463 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.523533 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.579442 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.580612 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.584095 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.585118 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vjvp4" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.587906 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7fd7273f-010c-48a0-9ecc-bd80e0bff239-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.588003 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fd7273f-010c-48a0-9ecc-bd80e0bff239-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.588042 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xkb9\" (UniqueName: \"kubernetes.io/projected/7fd7273f-010c-48a0-9ecc-bd80e0bff239-kube-api-access-8xkb9\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.588076 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.588130 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd7273f-010c-48a0-9ecc-bd80e0bff239-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.588168 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.588202 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.588267 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.594037 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.614896 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690497 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b310231-1c53-4831-aa00-e0f9597ce6e2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690592 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fd7273f-010c-48a0-9ecc-bd80e0bff239-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690652 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xkb9\" (UniqueName: \"kubernetes.io/projected/7fd7273f-010c-48a0-9ecc-bd80e0bff239-kube-api-access-8xkb9\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690728 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b310231-1c53-4831-aa00-e0f9597ce6e2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690764 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690793 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd7273f-010c-48a0-9ecc-bd80e0bff239-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690842 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690874 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690932 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b310231-1c53-4831-aa00-e0f9597ce6e2-kolla-config\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.690987 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjf4z\" (UniqueName: \"kubernetes.io/projected/1b310231-1c53-4831-aa00-e0f9597ce6e2-kube-api-access-hjf4z\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.691024 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.691092 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7fd7273f-010c-48a0-9ecc-bd80e0bff239-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.691123 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b310231-1c53-4831-aa00-e0f9597ce6e2-config-data\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.693457 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7fd7273f-010c-48a0-9ecc-bd80e0bff239-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.693617 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.693792 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.697564 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.699221 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd7273f-010c-48a0-9ecc-bd80e0bff239-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.709022 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7fd7273f-010c-48a0-9ecc-bd80e0bff239-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.709335 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fd7273f-010c-48a0-9ecc-bd80e0bff239-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.709458 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xkb9\" (UniqueName: \"kubernetes.io/projected/7fd7273f-010c-48a0-9ecc-bd80e0bff239-kube-api-access-8xkb9\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.719973 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7fd7273f-010c-48a0-9ecc-bd80e0bff239\") " pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.792956 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b310231-1c53-4831-aa00-e0f9597ce6e2-kolla-config\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.793018 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjf4z\" (UniqueName: \"kubernetes.io/projected/1b310231-1c53-4831-aa00-e0f9597ce6e2-kube-api-access-hjf4z\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.793068 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b310231-1c53-4831-aa00-e0f9597ce6e2-config-data\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.793105 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b310231-1c53-4831-aa00-e0f9597ce6e2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.793154 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b310231-1c53-4831-aa00-e0f9597ce6e2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.794040 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b310231-1c53-4831-aa00-e0f9597ce6e2-kolla-config\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.794486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b310231-1c53-4831-aa00-e0f9597ce6e2-config-data\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.802259 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b310231-1c53-4831-aa00-e0f9597ce6e2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.802821 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b310231-1c53-4831-aa00-e0f9597ce6e2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.812467 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjf4z\" (UniqueName: \"kubernetes.io/projected/1b310231-1c53-4831-aa00-e0f9597ce6e2-kube-api-access-hjf4z\") pod \"memcached-0\" (UID: \"1b310231-1c53-4831-aa00-e0f9597ce6e2\") " pod="openstack/memcached-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.867536 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 09 18:45:13 crc kubenswrapper[4750]: I0309 18:45:13.901463 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.172299 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.173559 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.177601 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-2c224" Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.187360 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.233732 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6vzz\" (UniqueName: \"kubernetes.io/projected/64ff0811-8670-4605-9ef6-383f43259ace-kube-api-access-z6vzz\") pod \"kube-state-metrics-0\" (UID: \"64ff0811-8670-4605-9ef6-383f43259ace\") " pod="openstack/kube-state-metrics-0" Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.334761 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6vzz\" (UniqueName: \"kubernetes.io/projected/64ff0811-8670-4605-9ef6-383f43259ace-kube-api-access-z6vzz\") pod \"kube-state-metrics-0\" (UID: \"64ff0811-8670-4605-9ef6-383f43259ace\") " pod="openstack/kube-state-metrics-0" Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.357130 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6vzz\" (UniqueName: \"kubernetes.io/projected/64ff0811-8670-4605-9ef6-383f43259ace-kube-api-access-z6vzz\") pod \"kube-state-metrics-0\" (UID: \"64ff0811-8670-4605-9ef6-383f43259ace\") " pod="openstack/kube-state-metrics-0" Mar 09 18:45:16 crc kubenswrapper[4750]: I0309 18:45:16.495548 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.486268 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.497766 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.501382 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.502305 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.503285 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-d8hlx" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.503447 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.504275 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.504424 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.509568 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.511204 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.514954 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.562684 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.562761 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.562788 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-config\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.562815 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.563133 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.563159 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.563184 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92888bd3-e3db-4127-97d2-37801bbbf1df-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.563214 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.563229 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9fng\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-kube-api-access-t9fng\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.563250 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664582 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664660 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664692 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-config\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664730 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664781 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664811 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664844 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92888bd3-e3db-4127-97d2-37801bbbf1df-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664881 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664900 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9fng\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-kube-api-access-t9fng\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.664921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.665794 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.666340 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.666988 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.673448 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.673680 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-config\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.676186 4750 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.676214 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c80612d5a9520b4a0639dd996533a4e971f0dc6e01d511a4a320a863e8837823/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.679215 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.680146 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92888bd3-e3db-4127-97d2-37801bbbf1df-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.691948 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.700355 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9fng\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-kube-api-access-t9fng\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.716408 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:17 crc kubenswrapper[4750]: I0309 18:45:17.822735 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.940010 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-n5qz8"] Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.941365 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.944150 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.944959 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.945211 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-687sd" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.948780 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6dgcc"] Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.950342 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.978796 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n5qz8"] Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.995554 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6dgcc"] Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.999338 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9fd826-3829-4379-9717-1b76ef376a39-combined-ca-bundle\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.999424 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wmqb\" (UniqueName: \"kubernetes.io/projected/5e9fd826-3829-4379-9717-1b76ef376a39-kube-api-access-4wmqb\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.999450 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-log-ovn\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.999474 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-run-ovn\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.999510 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9fd826-3829-4379-9717-1b76ef376a39-scripts\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.999533 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9fd826-3829-4379-9717-1b76ef376a39-ovn-controller-tls-certs\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:18 crc kubenswrapper[4750]: I0309 18:45:18.999564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-run\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.105925 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwpcg\" (UniqueName: \"kubernetes.io/projected/89841873-12eb-4add-bf8a-061b624c4b3a-kube-api-access-gwpcg\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106002 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wmqb\" (UniqueName: \"kubernetes.io/projected/5e9fd826-3829-4379-9717-1b76ef376a39-kube-api-access-4wmqb\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106032 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-log-ovn\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106056 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-run-ovn\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106089 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89841873-12eb-4add-bf8a-061b624c4b3a-scripts\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106131 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9fd826-3829-4379-9717-1b76ef376a39-scripts\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106170 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9fd826-3829-4379-9717-1b76ef376a39-ovn-controller-tls-certs\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106214 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-run\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106257 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-etc-ovs\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106301 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-log\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106336 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-lib\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106367 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9fd826-3829-4379-9717-1b76ef376a39-combined-ca-bundle\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.106393 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-run\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.107586 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-run\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.107749 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-run-ovn\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.107968 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e9fd826-3829-4379-9717-1b76ef376a39-var-log-ovn\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.110337 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9fd826-3829-4379-9717-1b76ef376a39-scripts\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.117606 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9fd826-3829-4379-9717-1b76ef376a39-ovn-controller-tls-certs\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.123510 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9fd826-3829-4379-9717-1b76ef376a39-combined-ca-bundle\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.144160 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wmqb\" (UniqueName: \"kubernetes.io/projected/5e9fd826-3829-4379-9717-1b76ef376a39-kube-api-access-4wmqb\") pod \"ovn-controller-n5qz8\" (UID: \"5e9fd826-3829-4379-9717-1b76ef376a39\") " pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.208229 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwpcg\" (UniqueName: \"kubernetes.io/projected/89841873-12eb-4add-bf8a-061b624c4b3a-kube-api-access-gwpcg\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.208942 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89841873-12eb-4add-bf8a-061b624c4b3a-scripts\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.213676 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89841873-12eb-4add-bf8a-061b624c4b3a-scripts\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.214028 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-etc-ovs\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.214790 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-etc-ovs\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.214961 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-log\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.215546 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-log\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.215663 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-lib\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.215731 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-run\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.215951 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-run\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.216129 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89841873-12eb-4add-bf8a-061b624c4b3a-var-lib\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.236340 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwpcg\" (UniqueName: \"kubernetes.io/projected/89841873-12eb-4add-bf8a-061b624c4b3a-kube-api-access-gwpcg\") pod \"ovn-controller-ovs-6dgcc\" (UID: \"89841873-12eb-4add-bf8a-061b624c4b3a\") " pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.277030 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:19 crc kubenswrapper[4750]: I0309 18:45:19.296774 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.541013 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.543695 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.548418 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.548990 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-spzxb" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.554493 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.554773 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.554838 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.556594 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.725762 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.725813 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.725857 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.725882 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx68m\" (UniqueName: \"kubernetes.io/projected/f059f39a-1c7f-46af-be90-4095105df733-kube-api-access-qx68m\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.725923 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f059f39a-1c7f-46af-be90-4095105df733-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.725945 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.725982 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f059f39a-1c7f-46af-be90-4095105df733-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.726009 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f059f39a-1c7f-46af-be90-4095105df733-config\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827208 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827263 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827303 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827326 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx68m\" (UniqueName: \"kubernetes.io/projected/f059f39a-1c7f-46af-be90-4095105df733-kube-api-access-qx68m\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827369 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f059f39a-1c7f-46af-be90-4095105df733-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827390 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827428 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f059f39a-1c7f-46af-be90-4095105df733-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.827469 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f059f39a-1c7f-46af-be90-4095105df733-config\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.828350 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f059f39a-1c7f-46af-be90-4095105df733-config\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.828492 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f059f39a-1c7f-46af-be90-4095105df733-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.829122 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f059f39a-1c7f-46af-be90-4095105df733-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.829255 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.834516 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.834548 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.842379 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f059f39a-1c7f-46af-be90-4095105df733-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.843207 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx68m\" (UniqueName: \"kubernetes.io/projected/f059f39a-1c7f-46af-be90-4095105df733-kube-api-access-qx68m\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.866829 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f059f39a-1c7f-46af-be90-4095105df733\") " pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:20 crc kubenswrapper[4750]: I0309 18:45:20.886298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.571545 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.573456 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.579546 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.579998 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.591556 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-hps2j" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.591795 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.608900 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707023 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707117 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707164 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707197 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxl5h\" (UniqueName: \"kubernetes.io/projected/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-kube-api-access-hxl5h\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707224 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707243 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-config\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707268 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.707337 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.808820 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.809080 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxl5h\" (UniqueName: \"kubernetes.io/projected/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-kube-api-access-hxl5h\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.809106 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.809129 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-config\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.809153 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.809226 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.809265 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.809318 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.810739 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.811031 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.811305 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.818202 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-config\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.819581 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.820223 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.825251 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.835087 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxl5h\" (UniqueName: \"kubernetes.io/projected/0dbcbd0d-a357-41da-9e5c-6672e4e326fe-kube-api-access-hxl5h\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.852355 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0dbcbd0d-a357-41da-9e5c-6672e4e326fe\") " pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:23 crc kubenswrapper[4750]: I0309 18:45:23.910954 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.537361 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.537426 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.537571 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.110:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mhkhz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-69554c7465-vpp94_openstack(2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.538979 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-69554c7465-vpp94" podUID="2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.574899 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.575031 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.575281 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.110:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vxq46,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-56d9f89cb5-6plnp_openstack(8cc465e5-9a2c-4774-b801-1f348194e086): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:24 crc kubenswrapper[4750]: E0309 18:45:24.576444 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" podUID="8cc465e5-9a2c-4774-b801-1f348194e086" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.013595 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-595d94d48f-57ntm"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.037706 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.137669 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxq46\" (UniqueName: \"kubernetes.io/projected/8cc465e5-9a2c-4774-b801-1f348194e086-kube-api-access-vxq46\") pod \"8cc465e5-9a2c-4774-b801-1f348194e086\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.137773 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-dns-svc\") pod \"8cc465e5-9a2c-4774-b801-1f348194e086\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.137942 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-config\") pod \"8cc465e5-9a2c-4774-b801-1f348194e086\" (UID: \"8cc465e5-9a2c-4774-b801-1f348194e086\") " Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.138533 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8cc465e5-9a2c-4774-b801-1f348194e086" (UID: "8cc465e5-9a2c-4774-b801-1f348194e086"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.138542 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-config" (OuterVolumeSpecName: "config") pod "8cc465e5-9a2c-4774-b801-1f348194e086" (UID: "8cc465e5-9a2c-4774-b801-1f348194e086"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.145456 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc465e5-9a2c-4774-b801-1f348194e086-kube-api-access-vxq46" (OuterVolumeSpecName: "kube-api-access-vxq46") pod "8cc465e5-9a2c-4774-b801-1f348194e086" (UID: "8cc465e5-9a2c-4774-b801-1f348194e086"). InnerVolumeSpecName "kube-api-access-vxq46". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.240750 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxq46\" (UniqueName: \"kubernetes.io/projected/8cc465e5-9a2c-4774-b801-1f348194e086-kube-api-access-vxq46\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.240794 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.240807 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc465e5-9a2c-4774-b801-1f348194e086-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.295658 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.444479 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-config\") pod \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.445102 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhkhz\" (UniqueName: \"kubernetes.io/projected/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-kube-api-access-mhkhz\") pod \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\" (UID: \"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3\") " Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.444920 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-config" (OuterVolumeSpecName: "config") pod "2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3" (UID: "2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.465566 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-kube-api-access-mhkhz" (OuterVolumeSpecName: "kube-api-access-mhkhz") pod "2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3" (UID: "2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3"). InnerVolumeSpecName "kube-api-access-mhkhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.500304 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fdf6c6f7-h2j9d"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.513261 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.519762 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.525805 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/notifications-rabbitmq-server-0"] Mar 09 18:45:25 crc kubenswrapper[4750]: W0309 18:45:25.534683 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod378d9b58_f830_4d61_b408_26668a301507.slice/crio-7b7a5f2b990282f78ad5069ff7216076ea76c26aca9c00a15b6a5a84c265577b WatchSource:0}: Error finding container 7b7a5f2b990282f78ad5069ff7216076ea76c26aca9c00a15b6a5a84c265577b: Status 404 returned error can't find the container with id 7b7a5f2b990282f78ad5069ff7216076ea76c26aca9c00a15b6a5a84c265577b Mar 09 18:45:25 crc kubenswrapper[4750]: W0309 18:45:25.535390 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3926781f_b03f_45a7_82e5_10abf0c16c1e.slice/crio-640d65e397a3e229d1145298fd34828858f5c51d206d454df8b0c8c95c73efe7 WatchSource:0}: Error finding container 640d65e397a3e229d1145298fd34828858f5c51d206d454df8b0c8c95c73efe7: Status 404 returned error can't find the container with id 640d65e397a3e229d1145298fd34828858f5c51d206d454df8b0c8c95c73efe7 Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.547376 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.547412 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhkhz\" (UniqueName: \"kubernetes.io/projected/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3-kube-api-access-mhkhz\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:25 crc kubenswrapper[4750]: W0309 18:45:25.550060 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe777a18_ce11_4d52_852d_b7bb3ffca188.slice/crio-5bef9afa553f1f15c64d54b03a3f7d33b77f752d87333433668ad50ba58e109d WatchSource:0}: Error finding container 5bef9afa553f1f15c64d54b03a3f7d33b77f752d87333433668ad50ba58e109d: Status 404 returned error can't find the container with id 5bef9afa553f1f15c64d54b03a3f7d33b77f752d87333433668ad50ba58e109d Mar 09 18:45:25 crc kubenswrapper[4750]: W0309 18:45:25.550466 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3236a6fb_c288_441c_96e5_7941f818b0af.slice/crio-b6ba2eed96cadb188db02a1152302b01183198beb4c60c2b24d08166f70fd40a WatchSource:0}: Error finding container b6ba2eed96cadb188db02a1152302b01183198beb4c60c2b24d08166f70fd40a: Status 404 returned error can't find the container with id b6ba2eed96cadb188db02a1152302b01183198beb4c60c2b24d08166f70fd40a Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.692571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3236a6fb-c288-441c-96e5-7941f818b0af","Type":"ContainerStarted","Data":"b6ba2eed96cadb188db02a1152302b01183198beb4c60c2b24d08166f70fd40a"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.700055 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"3926781f-b03f-45a7-82e5-10abf0c16c1e","Type":"ContainerStarted","Data":"640d65e397a3e229d1145298fd34828858f5c51d206d454df8b0c8c95c73efe7"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.705768 4750 generic.go:334] "Generic (PLEG): container finished" podID="e8ab0211-a492-45f9-9606-08706461761d" containerID="cc0b2db5e27692056731cd281a566068c49541e0931b36c5c30539f34b00a05a" exitCode=0 Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.705822 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" event={"ID":"e8ab0211-a492-45f9-9606-08706461761d","Type":"ContainerDied","Data":"cc0b2db5e27692056731cd281a566068c49541e0931b36c5c30539f34b00a05a"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.705874 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" event={"ID":"e8ab0211-a492-45f9-9606-08706461761d","Type":"ContainerStarted","Data":"abb1579db2a267921f130faefa34e53728a4bf58b06726024c7ed1df19a774b6"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.707738 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" event={"ID":"8cc465e5-9a2c-4774-b801-1f348194e086","Type":"ContainerDied","Data":"4d3a82cfe7ec846f198b2ad120d8980a11f4fa888ac805cc7570ad9125295259"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.707796 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d9f89cb5-6plnp" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.710160 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69554c7465-vpp94" event={"ID":"2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3","Type":"ContainerDied","Data":"a08a31fe30417068949185c62b5310be40661dd879cab24bdfe821c8ceee724c"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.710218 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69554c7465-vpp94" Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.712987 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"378d9b58-f830-4d61-b408-26668a301507","Type":"ContainerStarted","Data":"7b7a5f2b990282f78ad5069ff7216076ea76c26aca9c00a15b6a5a84c265577b"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.714562 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" event={"ID":"be777a18-ce11-4d52-852d-b7bb3ffca188","Type":"ContainerStarted","Data":"5bef9afa553f1f15c64d54b03a3f7d33b77f752d87333433668ad50ba58e109d"} Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.796606 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56d9f89cb5-6plnp"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.804580 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56d9f89cb5-6plnp"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.835331 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69554c7465-vpp94"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.844121 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69554c7465-vpp94"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.975253 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n5qz8"] Mar 09 18:45:25 crc kubenswrapper[4750]: I0309 18:45:25.996669 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.011341 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:45:26 crc kubenswrapper[4750]: W0309 18:45:26.019405 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64ff0811_8670_4605_9ef6_383f43259ace.slice/crio-9eb1fbbffa42af84ada906b4a3bb797f2f2794dfe009274acd26470d2aec7191 WatchSource:0}: Error finding container 9eb1fbbffa42af84ada906b4a3bb797f2f2794dfe009274acd26470d2aec7191: Status 404 returned error can't find the container with id 9eb1fbbffa42af84ada906b4a3bb797f2f2794dfe009274acd26470d2aec7191 Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.024021 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:45:26 crc kubenswrapper[4750]: W0309 18:45:26.053366 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92888bd3_e3db_4127_97d2_37801bbbf1df.slice/crio-faad1e19f72a60152ea7f79c423bfeb6880334dde764c0e096c2618519521912 WatchSource:0}: Error finding container faad1e19f72a60152ea7f79c423bfeb6880334dde764c0e096c2618519521912: Status 404 returned error can't find the container with id faad1e19f72a60152ea7f79c423bfeb6880334dde764c0e096c2618519521912 Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.170240 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.201808 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b67658d95-pfl2f"] Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.224184 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.238904 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 09 18:45:26 crc kubenswrapper[4750]: W0309 18:45:26.247354 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fd7273f_010c_48a0_9ecc_bd80e0bff239.slice/crio-090e372fc9089181dd05d1d352b4355188dad08292d4dc8e018a5a9d15f82fc6 WatchSource:0}: Error finding container 090e372fc9089181dd05d1d352b4355188dad08292d4dc8e018a5a9d15f82fc6: Status 404 returned error can't find the container with id 090e372fc9089181dd05d1d352b4355188dad08292d4dc8e018a5a9d15f82fc6 Mar 09 18:45:26 crc kubenswrapper[4750]: W0309 18:45:26.250442 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc48f9535_ad2d_4cac_a39c_043f1d25859a.slice/crio-5d41ef0fae47cb3c388149e6a55ebe76e8e7da3052a544209f06599cdd9cfd89 WatchSource:0}: Error finding container 5d41ef0fae47cb3c388149e6a55ebe76e8e7da3052a544209f06599cdd9cfd89: Status 404 returned error can't find the container with id 5d41ef0fae47cb3c388149e6a55ebe76e8e7da3052a544209f06599cdd9cfd89 Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.262651 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6dgcc"] Mar 09 18:45:26 crc kubenswrapper[4750]: W0309 18:45:26.290721 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89841873_12eb_4add_bf8a_061b624c4b3a.slice/crio-b92ae75d41947850718bd448b2da29a2a9fa458ddd6c039d284fd77281331b29 WatchSource:0}: Error finding container b92ae75d41947850718bd448b2da29a2a9fa458ddd6c039d284fd77281331b29: Status 404 returned error can't find the container with id b92ae75d41947850718bd448b2da29a2a9fa458ddd6c039d284fd77281331b29 Mar 09 18:45:26 crc kubenswrapper[4750]: E0309 18:45:26.699526 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc48f9535_ad2d_4cac_a39c_043f1d25859a.slice/crio-7636b075edc29bbe082289be06581ffc3d2affbbd797f64f5fdaf5de2b1479dd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc48f9535_ad2d_4cac_a39c_043f1d25859a.slice/crio-conmon-7636b075edc29bbe082289be06581ffc3d2affbbd797f64f5fdaf5de2b1479dd.scope\": RecentStats: unable to find data in memory cache]" Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.738946 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7fd7273f-010c-48a0-9ecc-bd80e0bff239","Type":"ContainerStarted","Data":"090e372fc9089181dd05d1d352b4355188dad08292d4dc8e018a5a9d15f82fc6"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.759322 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" event={"ID":"e8ab0211-a492-45f9-9606-08706461761d","Type":"ContainerStarted","Data":"e2a62efbca093ecd19a98c3d88d2dbb4b29f57dbe52ac80b0bc465b40db13bd6"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.759401 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.778132 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"302428aa-7eba-4f8b-8868-1e8e883d38c2","Type":"ContainerStarted","Data":"cfd0f5ea2b7cb5e55e9b18ad61761e81853c6c6d504564bd8ac693a62789e463"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.787185 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n5qz8" event={"ID":"5e9fd826-3829-4379-9717-1b76ef376a39","Type":"ContainerStarted","Data":"e493170cb0988f368f97b17f2b24854bf5266741492e659d6d3cf3f0e1868c59"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.793243 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerStarted","Data":"faad1e19f72a60152ea7f79c423bfeb6880334dde764c0e096c2618519521912"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.802767 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"1b310231-1c53-4831-aa00-e0f9597ce6e2","Type":"ContainerStarted","Data":"81f775bcbe0cf42bc18ea7e5e71595e0b862b94191796f0319437104b87dcace"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.806513 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.812840 4750 generic.go:334] "Generic (PLEG): container finished" podID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerID="7636b075edc29bbe082289be06581ffc3d2affbbd797f64f5fdaf5de2b1479dd" exitCode=0 Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.812908 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" event={"ID":"c48f9535-ad2d-4cac-a39c-043f1d25859a","Type":"ContainerDied","Data":"7636b075edc29bbe082289be06581ffc3d2affbbd797f64f5fdaf5de2b1479dd"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.812943 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" event={"ID":"c48f9535-ad2d-4cac-a39c-043f1d25859a","Type":"ContainerStarted","Data":"5d41ef0fae47cb3c388149e6a55ebe76e8e7da3052a544209f06599cdd9cfd89"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.822285 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6dgcc" event={"ID":"89841873-12eb-4add-bf8a-061b624c4b3a","Type":"ContainerStarted","Data":"b92ae75d41947850718bd448b2da29a2a9fa458ddd6c039d284fd77281331b29"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.827070 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" podStartSLOduration=17.737306232999998 podStartE2EDuration="17.827044306s" podCreationTimestamp="2026-03-09 18:45:09 +0000 UTC" firstStartedPulling="2026-03-09 18:45:25.051120157 +0000 UTC m=+1206.393592555" lastFinishedPulling="2026-03-09 18:45:25.14085823 +0000 UTC m=+1206.483330628" observedRunningTime="2026-03-09 18:45:26.794472854 +0000 UTC m=+1208.136945242" watchObservedRunningTime="2026-03-09 18:45:26.827044306 +0000 UTC m=+1208.169516704" Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.827327 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"64ff0811-8670-4605-9ef6-383f43259ace","Type":"ContainerStarted","Data":"9eb1fbbffa42af84ada906b4a3bb797f2f2794dfe009274acd26470d2aec7191"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.830691 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f059f39a-1c7f-46af-be90-4095105df733","Type":"ContainerStarted","Data":"7bce152ce66de96b23ba7b7cf09632dfbb0b881fd0fc147348679b3004f47b79"} Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.841191 4750 generic.go:334] "Generic (PLEG): container finished" podID="be777a18-ce11-4d52-852d-b7bb3ffca188" containerID="25a55b14a579297a3ec0a1cfc76a17ceeeeb6e6bffee5577d83663e21472103c" exitCode=0 Mar 09 18:45:26 crc kubenswrapper[4750]: I0309 18:45:26.841264 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" event={"ID":"be777a18-ce11-4d52-852d-b7bb3ffca188","Type":"ContainerDied","Data":"25a55b14a579297a3ec0a1cfc76a17ceeeeb6e6bffee5577d83663e21472103c"} Mar 09 18:45:27 crc kubenswrapper[4750]: W0309 18:45:27.225256 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dbcbd0d_a357_41da_9e5c_6672e4e326fe.slice/crio-2b6f001384a1608802be71d409a321d6dc339523760cc728cf7d37810e21b0c5 WatchSource:0}: Error finding container 2b6f001384a1608802be71d409a321d6dc339523760cc728cf7d37810e21b0c5: Status 404 returned error can't find the container with id 2b6f001384a1608802be71d409a321d6dc339523760cc728cf7d37810e21b0c5 Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.389369 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3" path="/var/lib/kubelet/pods/2fc64f46-40c3-435b-8c8f-ae5e4ee0bfa3/volumes" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.389809 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc465e5-9a2c-4774-b801-1f348194e086" path="/var/lib/kubelet/pods/8cc465e5-9a2c-4774-b801-1f348194e086/volumes" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.438334 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.608212 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-dns-svc\") pod \"be777a18-ce11-4d52-852d-b7bb3ffca188\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.608282 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-config\") pod \"be777a18-ce11-4d52-852d-b7bb3ffca188\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.608385 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkxnw\" (UniqueName: \"kubernetes.io/projected/be777a18-ce11-4d52-852d-b7bb3ffca188-kube-api-access-bkxnw\") pod \"be777a18-ce11-4d52-852d-b7bb3ffca188\" (UID: \"be777a18-ce11-4d52-852d-b7bb3ffca188\") " Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.615037 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be777a18-ce11-4d52-852d-b7bb3ffca188-kube-api-access-bkxnw" (OuterVolumeSpecName: "kube-api-access-bkxnw") pod "be777a18-ce11-4d52-852d-b7bb3ffca188" (UID: "be777a18-ce11-4d52-852d-b7bb3ffca188"). InnerVolumeSpecName "kube-api-access-bkxnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.632277 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "be777a18-ce11-4d52-852d-b7bb3ffca188" (UID: "be777a18-ce11-4d52-852d-b7bb3ffca188"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.633642 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-config" (OuterVolumeSpecName: "config") pod "be777a18-ce11-4d52-852d-b7bb3ffca188" (UID: "be777a18-ce11-4d52-852d-b7bb3ffca188"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.710571 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.710608 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkxnw\" (UniqueName: \"kubernetes.io/projected/be777a18-ce11-4d52-852d-b7bb3ffca188-kube-api-access-bkxnw\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.710621 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be777a18-ce11-4d52-852d-b7bb3ffca188-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.859462 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" event={"ID":"be777a18-ce11-4d52-852d-b7bb3ffca188","Type":"ContainerDied","Data":"5bef9afa553f1f15c64d54b03a3f7d33b77f752d87333433668ad50ba58e109d"} Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.859549 4750 scope.go:117] "RemoveContainer" containerID="25a55b14a579297a3ec0a1cfc76a17ceeeeb6e6bffee5577d83663e21472103c" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.859740 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdf6c6f7-h2j9d" Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.867622 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0dbcbd0d-a357-41da-9e5c-6672e4e326fe","Type":"ContainerStarted","Data":"2b6f001384a1608802be71d409a321d6dc339523760cc728cf7d37810e21b0c5"} Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.950706 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fdf6c6f7-h2j9d"] Mar 09 18:45:27 crc kubenswrapper[4750]: I0309 18:45:27.956903 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fdf6c6f7-h2j9d"] Mar 09 18:45:29 crc kubenswrapper[4750]: I0309 18:45:29.393719 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be777a18-ce11-4d52-852d-b7bb3ffca188" path="/var/lib/kubelet/pods/be777a18-ce11-4d52-852d-b7bb3ffca188/volumes" Mar 09 18:45:34 crc kubenswrapper[4750]: I0309 18:45:34.953904 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:45:35 crc kubenswrapper[4750]: I0309 18:45:35.019821 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b67658d95-pfl2f"] Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.625103 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ovn-base:watcher_latest" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.625436 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ovn-base:watcher_latest" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.625586 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:38.102.83.110:5001/podified-master-centos10/openstack-ovn-base:watcher_latest,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc5h9bh5f6hbbh8bh5c8h5dhf7h5c4hffh589hcch56dh5cbh65bh646h575hc7h556h56dh68h59fh684h4h65ch5c9h75h9dh66dh6fhf5h54q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gwpcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-6dgcc_openstack(89841873-12eb-4add-bf8a-061b624c4b3a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.627995 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-6dgcc" podUID="89841873-12eb-4add-bf8a-061b624c4b3a" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.696812 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.697412 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.697655 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:38.102.83.110:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r78lr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod notifications-rabbitmq-server-0_openstack(3926781f-b03f-45a7-82e5-10abf0c16c1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.699260 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/notifications-rabbitmq-server-0" podUID="3926781f-b03f-45a7-82e5-10abf0c16c1e" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.972891 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-ovn-base:watcher_latest\\\"\"" pod="openstack/ovn-controller-ovs-6dgcc" podUID="89841873-12eb-4add-bf8a-061b624c4b3a" Mar 09 18:45:37 crc kubenswrapper[4750]: E0309 18:45:37.973029 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest\\\"\"" pod="openstack/notifications-rabbitmq-server-0" podUID="3926781f-b03f-45a7-82e5-10abf0c16c1e" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.184377 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.186088 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.186478 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-89tht,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(302428aa-7eba-4f8b-8868-1e8e883d38c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.188037 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="302428aa-7eba-4f8b-8868-1e8e883d38c2" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.248953 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.249502 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.250018 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8xkb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(7fd7273f-010c-48a0-9ecc-bd80e0bff239): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.251536 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="7fd7273f-010c-48a0-9ecc-bd80e0bff239" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.628668 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.628739 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.629536 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:38.102.83.110:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n59dh7h57fh57dhb8h56fh65ch66bh77h5ch557hc8h548h685h9ch684h58ch5dbh5b7h585hd5h5d9h567h574h5ddh598h8fh694h647hd8hfh64bq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qx68m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(f059f39a-1c7f-46af-be90-4095105df733): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.826014 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ovn-controller:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.826083 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ovn-controller:watcher_latest" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.826316 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:38.102.83.110:5001/podified-master-centos10/openstack-ovn-controller:watcher_latest,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc5h9bh5f6hbbh8bh5c8h5dhf7h5c4hffh589hcch56dh5cbh65bh646h575hc7h556h56dh68h59fh684h4h65ch5c9h75h9dh66dh6fhf5h54q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wmqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-n5qz8_openstack(5e9fd826-3829-4379-9717-1b76ef376a39): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:45:40 crc kubenswrapper[4750]: E0309 18:45:40.827557 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-n5qz8" podUID="5e9fd826-3829-4379-9717-1b76ef376a39" Mar 09 18:45:41 crc kubenswrapper[4750]: E0309 18:45:41.011610 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest\\\"\"" pod="openstack/openstack-galera-0" podUID="302428aa-7eba-4f8b-8868-1e8e883d38c2" Mar 09 18:45:41 crc kubenswrapper[4750]: E0309 18:45:41.013325 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-mariadb:watcher_latest\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="7fd7273f-010c-48a0-9ecc-bd80e0bff239" Mar 09 18:45:41 crc kubenswrapper[4750]: E0309 18:45:41.020466 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-ovn-controller:watcher_latest\\\"\"" pod="openstack/ovn-controller-n5qz8" podUID="5e9fd826-3829-4379-9717-1b76ef376a39" Mar 09 18:45:42 crc kubenswrapper[4750]: E0309 18:45:42.034402 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 09 18:45:42 crc kubenswrapper[4750]: E0309 18:45:42.034474 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 09 18:45:42 crc kubenswrapper[4750]: E0309 18:45:42.034707 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z6vzz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(64ff0811-8670-4605-9ef6-383f43259ace): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 18:45:42 crc kubenswrapper[4750]: E0309 18:45:42.036385 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="64ff0811-8670-4605-9ef6-383f43259ace" Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.027730 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" event={"ID":"c48f9535-ad2d-4cac-a39c-043f1d25859a","Type":"ContainerStarted","Data":"c21f918402e6c0132733e6ff0642c4d2bde82e76ef8b5a09f80d6d1d395c3db8"} Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.027917 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" podUID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerName="dnsmasq-dns" containerID="cri-o://c21f918402e6c0132733e6ff0642c4d2bde82e76ef8b5a09f80d6d1d395c3db8" gracePeriod=10 Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.028156 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.029411 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0dbcbd0d-a357-41da-9e5c-6672e4e326fe","Type":"ContainerStarted","Data":"4173f105d2aa56d1ebb43382862b4463eead28186f61ca6183a45f689d0b2422"} Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.032045 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"1b310231-1c53-4831-aa00-e0f9597ce6e2","Type":"ContainerStarted","Data":"31c948ff2eda6c5fd8900e499d3dc8e6b5488c15205809f87e8f42b4669f5fd4"} Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.032070 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 09 18:45:43 crc kubenswrapper[4750]: E0309 18:45:43.033313 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="64ff0811-8670-4605-9ef6-383f43259ace" Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.050968 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" podStartSLOduration=34.050947901 podStartE2EDuration="34.050947901s" podCreationTimestamp="2026-03-09 18:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:45:43.045252707 +0000 UTC m=+1224.387725105" watchObservedRunningTime="2026-03-09 18:45:43.050947901 +0000 UTC m=+1224.393420299" Mar 09 18:45:43 crc kubenswrapper[4750]: I0309 18:45:43.079082 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=15.062726861 podStartE2EDuration="30.079041753s" podCreationTimestamp="2026-03-09 18:45:13 +0000 UTC" firstStartedPulling="2026-03-09 18:45:25.985664789 +0000 UTC m=+1207.328137187" lastFinishedPulling="2026-03-09 18:45:41.001979681 +0000 UTC m=+1222.344452079" observedRunningTime="2026-03-09 18:45:43.063348588 +0000 UTC m=+1224.405820996" watchObservedRunningTime="2026-03-09 18:45:43.079041753 +0000 UTC m=+1224.421514151" Mar 09 18:45:44 crc kubenswrapper[4750]: I0309 18:45:44.055544 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"378d9b58-f830-4d61-b408-26668a301507","Type":"ContainerStarted","Data":"bbc618a451438e977e10a76341e34bb8d2c1f770d1a8c8ffc94ff70b158dfb10"} Mar 09 18:45:44 crc kubenswrapper[4750]: I0309 18:45:44.058875 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3236a6fb-c288-441c-96e5-7941f818b0af","Type":"ContainerStarted","Data":"4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143"} Mar 09 18:45:44 crc kubenswrapper[4750]: I0309 18:45:44.064161 4750 generic.go:334] "Generic (PLEG): container finished" podID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerID="c21f918402e6c0132733e6ff0642c4d2bde82e76ef8b5a09f80d6d1d395c3db8" exitCode=0 Mar 09 18:45:44 crc kubenswrapper[4750]: I0309 18:45:44.064237 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" event={"ID":"c48f9535-ad2d-4cac-a39c-043f1d25859a","Type":"ContainerDied","Data":"c21f918402e6c0132733e6ff0642c4d2bde82e76ef8b5a09f80d6d1d395c3db8"} Mar 09 18:45:45 crc kubenswrapper[4750]: I0309 18:45:45.078355 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerStarted","Data":"1624d9c60026bcd3e01d587842fe8b34482e04207adfcf064ea1215fd2120205"} Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.495144 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.625813 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-dns-svc\") pod \"c48f9535-ad2d-4cac-a39c-043f1d25859a\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.626312 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-config\") pod \"c48f9535-ad2d-4cac-a39c-043f1d25859a\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.626477 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b9w7\" (UniqueName: \"kubernetes.io/projected/c48f9535-ad2d-4cac-a39c-043f1d25859a-kube-api-access-6b9w7\") pod \"c48f9535-ad2d-4cac-a39c-043f1d25859a\" (UID: \"c48f9535-ad2d-4cac-a39c-043f1d25859a\") " Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.634591 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48f9535-ad2d-4cac-a39c-043f1d25859a-kube-api-access-6b9w7" (OuterVolumeSpecName: "kube-api-access-6b9w7") pod "c48f9535-ad2d-4cac-a39c-043f1d25859a" (UID: "c48f9535-ad2d-4cac-a39c-043f1d25859a"). InnerVolumeSpecName "kube-api-access-6b9w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.682458 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-config" (OuterVolumeSpecName: "config") pod "c48f9535-ad2d-4cac-a39c-043f1d25859a" (UID: "c48f9535-ad2d-4cac-a39c-043f1d25859a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.687699 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c48f9535-ad2d-4cac-a39c-043f1d25859a" (UID: "c48f9535-ad2d-4cac-a39c-043f1d25859a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.728966 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.729175 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48f9535-ad2d-4cac-a39c-043f1d25859a-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:46 crc kubenswrapper[4750]: I0309 18:45:46.729239 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b9w7\" (UniqueName: \"kubernetes.io/projected/c48f9535-ad2d-4cac-a39c-043f1d25859a-kube-api-access-6b9w7\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:46 crc kubenswrapper[4750]: E0309 18:45:46.750555 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="f059f39a-1c7f-46af-be90-4095105df733" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.097648 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f059f39a-1c7f-46af-be90-4095105df733","Type":"ContainerStarted","Data":"fb94f716bfa77f5cf6272880664ee7ae28d91d56faf3dae7f20332b08d29c4e0"} Mar 09 18:45:47 crc kubenswrapper[4750]: E0309 18:45:47.099654 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="f059f39a-1c7f-46af-be90-4095105df733" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.103375 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" event={"ID":"c48f9535-ad2d-4cac-a39c-043f1d25859a","Type":"ContainerDied","Data":"5d41ef0fae47cb3c388149e6a55ebe76e8e7da3052a544209f06599cdd9cfd89"} Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.103479 4750 scope.go:117] "RemoveContainer" containerID="c21f918402e6c0132733e6ff0642c4d2bde82e76ef8b5a09f80d6d1d395c3db8" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.103413 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b67658d95-pfl2f" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.107973 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0dbcbd0d-a357-41da-9e5c-6672e4e326fe","Type":"ContainerStarted","Data":"9ca538a8e7b61377387b2b4dc50aa1ea2d5c09e8073ef91dde7a9ab513043465"} Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.142739 4750 scope.go:117] "RemoveContainer" containerID="7636b075edc29bbe082289be06581ffc3d2affbbd797f64f5fdaf5de2b1479dd" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.180508 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.945756351 podStartE2EDuration="25.180413877s" podCreationTimestamp="2026-03-09 18:45:22 +0000 UTC" firstStartedPulling="2026-03-09 18:45:27.259434507 +0000 UTC m=+1208.601906905" lastFinishedPulling="2026-03-09 18:45:46.494092033 +0000 UTC m=+1227.836564431" observedRunningTime="2026-03-09 18:45:47.165149754 +0000 UTC m=+1228.507622152" watchObservedRunningTime="2026-03-09 18:45:47.180413877 +0000 UTC m=+1228.522886275" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.220774 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b67658d95-pfl2f"] Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.231062 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b67658d95-pfl2f"] Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.389863 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48f9535-ad2d-4cac-a39c-043f1d25859a" path="/var/lib/kubelet/pods/c48f9535-ad2d-4cac-a39c-043f1d25859a/volumes" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.911589 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:47 crc kubenswrapper[4750]: I0309 18:45:47.953021 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.118825 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:48 crc kubenswrapper[4750]: E0309 18:45:48.120317 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="f059f39a-1c7f-46af-be90-4095105df733" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.181835 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.474484 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66fb8f999f-knd9k"] Mar 09 18:45:48 crc kubenswrapper[4750]: E0309 18:45:48.474940 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be777a18-ce11-4d52-852d-b7bb3ffca188" containerName="init" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.474961 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="be777a18-ce11-4d52-852d-b7bb3ffca188" containerName="init" Mar 09 18:45:48 crc kubenswrapper[4750]: E0309 18:45:48.474984 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerName="dnsmasq-dns" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.474997 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerName="dnsmasq-dns" Mar 09 18:45:48 crc kubenswrapper[4750]: E0309 18:45:48.475022 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerName="init" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.475030 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerName="init" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.475243 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48f9535-ad2d-4cac-a39c-043f1d25859a" containerName="dnsmasq-dns" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.475262 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="be777a18-ce11-4d52-852d-b7bb3ffca188" containerName="init" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.476380 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.478368 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.488274 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66fb8f999f-knd9k"] Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.535993 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-nz6xk"] Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.537676 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.547095 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nz6xk"] Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.547358 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573483 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbqm4\" (UniqueName: \"kubernetes.io/projected/4a73090f-b06c-4eb9-bb78-6b91f20984d0-kube-api-access-nbqm4\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573538 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ab758bdb-9d1f-4941-8c95-84cc4c867f09-ovs-rundir\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573584 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x7s6\" (UniqueName: \"kubernetes.io/projected/ab758bdb-9d1f-4941-8c95-84cc4c867f09-kube-api-access-9x7s6\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573652 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-dns-svc\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573689 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab758bdb-9d1f-4941-8c95-84cc4c867f09-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573720 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ab758bdb-9d1f-4941-8c95-84cc4c867f09-ovn-rundir\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573753 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab758bdb-9d1f-4941-8c95-84cc4c867f09-combined-ca-bundle\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573780 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-config\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.573808 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-ovsdbserver-sb\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.574089 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab758bdb-9d1f-4941-8c95-84cc4c867f09-config\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.675858 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab758bdb-9d1f-4941-8c95-84cc4c867f09-config\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676287 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbqm4\" (UniqueName: \"kubernetes.io/projected/4a73090f-b06c-4eb9-bb78-6b91f20984d0-kube-api-access-nbqm4\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676315 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ab758bdb-9d1f-4941-8c95-84cc4c867f09-ovs-rundir\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676352 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x7s6\" (UniqueName: \"kubernetes.io/projected/ab758bdb-9d1f-4941-8c95-84cc4c867f09-kube-api-access-9x7s6\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676389 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-dns-svc\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676416 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab758bdb-9d1f-4941-8c95-84cc4c867f09-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676443 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ab758bdb-9d1f-4941-8c95-84cc4c867f09-ovn-rundir\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676467 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab758bdb-9d1f-4941-8c95-84cc4c867f09-combined-ca-bundle\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676486 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-config\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.676506 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-ovsdbserver-sb\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.677342 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-ovsdbserver-sb\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.678010 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-dns-svc\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.678712 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab758bdb-9d1f-4941-8c95-84cc4c867f09-config\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.679124 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ab758bdb-9d1f-4941-8c95-84cc4c867f09-ovs-rundir\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.679280 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ab758bdb-9d1f-4941-8c95-84cc4c867f09-ovn-rundir\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.679991 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-config\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.689685 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab758bdb-9d1f-4941-8c95-84cc4c867f09-combined-ca-bundle\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.691519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab758bdb-9d1f-4941-8c95-84cc4c867f09-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.704538 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x7s6\" (UniqueName: \"kubernetes.io/projected/ab758bdb-9d1f-4941-8c95-84cc4c867f09-kube-api-access-9x7s6\") pod \"ovn-controller-metrics-nz6xk\" (UID: \"ab758bdb-9d1f-4941-8c95-84cc4c867f09\") " pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.706429 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbqm4\" (UniqueName: \"kubernetes.io/projected/4a73090f-b06c-4eb9-bb78-6b91f20984d0-kube-api-access-nbqm4\") pod \"dnsmasq-dns-66fb8f999f-knd9k\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.793805 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.820932 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fb8f999f-knd9k"] Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.860613 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cb4b9444f-7dhb2"] Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.872085 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.874521 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cb4b9444f-7dhb2"] Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.877246 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.882708 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nz6xk" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.884909 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-sb\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.885135 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-nb\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.885214 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-config\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.885301 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-dns-svc\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.885508 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztb4q\" (UniqueName: \"kubernetes.io/projected/a90672b1-a38d-4f7a-b770-22fab99c5474-kube-api-access-ztb4q\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.906800 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.987258 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-dns-svc\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.987354 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztb4q\" (UniqueName: \"kubernetes.io/projected/a90672b1-a38d-4f7a-b770-22fab99c5474-kube-api-access-ztb4q\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.987439 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-sb\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.987555 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-nb\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.987603 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-config\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.988477 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-dns-svc\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.988728 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-config\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.989274 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-sb\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:48 crc kubenswrapper[4750]: I0309 18:45:48.989476 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-nb\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:49 crc kubenswrapper[4750]: I0309 18:45:49.023014 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztb4q\" (UniqueName: \"kubernetes.io/projected/a90672b1-a38d-4f7a-b770-22fab99c5474-kube-api-access-ztb4q\") pod \"dnsmasq-dns-6cb4b9444f-7dhb2\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:49 crc kubenswrapper[4750]: I0309 18:45:49.175332 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fb8f999f-knd9k"] Mar 09 18:45:49 crc kubenswrapper[4750]: I0309 18:45:49.234317 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:49 crc kubenswrapper[4750]: I0309 18:45:49.252179 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nz6xk"] Mar 09 18:45:49 crc kubenswrapper[4750]: I0309 18:45:49.739037 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cb4b9444f-7dhb2"] Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.134378 4750 generic.go:334] "Generic (PLEG): container finished" podID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerID="d768357ba96d4a2d8d721908bf89b7b1539d671d06c5c42d87f18482da335dcf" exitCode=0 Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.134460 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" event={"ID":"a90672b1-a38d-4f7a-b770-22fab99c5474","Type":"ContainerDied","Data":"d768357ba96d4a2d8d721908bf89b7b1539d671d06c5c42d87f18482da335dcf"} Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.134537 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" event={"ID":"a90672b1-a38d-4f7a-b770-22fab99c5474","Type":"ContainerStarted","Data":"f0e382f70d3048534c32f268cb080a800b0dbe622de0aca60bb89138b6f6f8c4"} Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.136416 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nz6xk" event={"ID":"ab758bdb-9d1f-4941-8c95-84cc4c867f09","Type":"ContainerStarted","Data":"48524d9104f470a9ba22120e0065c3472732bdd8d4933f57df7a0c9a6b75708e"} Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.136467 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nz6xk" event={"ID":"ab758bdb-9d1f-4941-8c95-84cc4c867f09","Type":"ContainerStarted","Data":"3e576aee5598d5301c61b2e43111195616f7b445a5741e46d35b6f644e99eed0"} Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.138330 4750 generic.go:334] "Generic (PLEG): container finished" podID="4a73090f-b06c-4eb9-bb78-6b91f20984d0" containerID="0f141abf513f863c0868b477e50d6edbd5c50a2d689376924bf5dc94be4df060" exitCode=0 Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.138962 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" event={"ID":"4a73090f-b06c-4eb9-bb78-6b91f20984d0","Type":"ContainerDied","Data":"0f141abf513f863c0868b477e50d6edbd5c50a2d689376924bf5dc94be4df060"} Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.138997 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" event={"ID":"4a73090f-b06c-4eb9-bb78-6b91f20984d0","Type":"ContainerStarted","Data":"ee993398a05408a63daa3f1e6405b38d6154baed05e7540e11e8805f6eb37264"} Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.187601 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-nz6xk" podStartSLOduration=2.187573671 podStartE2EDuration="2.187573671s" podCreationTimestamp="2026-03-09 18:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:45:50.177305233 +0000 UTC m=+1231.519777631" watchObservedRunningTime="2026-03-09 18:45:50.187573671 +0000 UTC m=+1231.530046079" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.545472 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.623439 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-ovsdbserver-sb\") pod \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.623524 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbqm4\" (UniqueName: \"kubernetes.io/projected/4a73090f-b06c-4eb9-bb78-6b91f20984d0-kube-api-access-nbqm4\") pod \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.623576 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-config\") pod \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.625554 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-dns-svc\") pod \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\" (UID: \"4a73090f-b06c-4eb9-bb78-6b91f20984d0\") " Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.638234 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a73090f-b06c-4eb9-bb78-6b91f20984d0-kube-api-access-nbqm4" (OuterVolumeSpecName: "kube-api-access-nbqm4") pod "4a73090f-b06c-4eb9-bb78-6b91f20984d0" (UID: "4a73090f-b06c-4eb9-bb78-6b91f20984d0"). InnerVolumeSpecName "kube-api-access-nbqm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.653137 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4a73090f-b06c-4eb9-bb78-6b91f20984d0" (UID: "4a73090f-b06c-4eb9-bb78-6b91f20984d0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.676319 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-config" (OuterVolumeSpecName: "config") pod "4a73090f-b06c-4eb9-bb78-6b91f20984d0" (UID: "4a73090f-b06c-4eb9-bb78-6b91f20984d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.685462 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4a73090f-b06c-4eb9-bb78-6b91f20984d0" (UID: "4a73090f-b06c-4eb9-bb78-6b91f20984d0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.729293 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.729324 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbqm4\" (UniqueName: \"kubernetes.io/projected/4a73090f-b06c-4eb9-bb78-6b91f20984d0-kube-api-access-nbqm4\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.729338 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:50 crc kubenswrapper[4750]: I0309 18:45:50.729353 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a73090f-b06c-4eb9-bb78-6b91f20984d0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.151482 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" event={"ID":"a90672b1-a38d-4f7a-b770-22fab99c5474","Type":"ContainerStarted","Data":"c2e32c99687231ab22b644db43f701f17bb17861132261104abb1a5927b7a0a7"} Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.153040 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.154469 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.154495 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fb8f999f-knd9k" event={"ID":"4a73090f-b06c-4eb9-bb78-6b91f20984d0","Type":"ContainerDied","Data":"ee993398a05408a63daa3f1e6405b38d6154baed05e7540e11e8805f6eb37264"} Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.154572 4750 scope.go:117] "RemoveContainer" containerID="0f141abf513f863c0868b477e50d6edbd5c50a2d689376924bf5dc94be4df060" Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.156988 4750 generic.go:334] "Generic (PLEG): container finished" podID="89841873-12eb-4add-bf8a-061b624c4b3a" containerID="7eadb2058d2aa243856cd00ec267e67e2d158e64199a746e3460ff8c4b192b00" exitCode=0 Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.157036 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6dgcc" event={"ID":"89841873-12eb-4add-bf8a-061b624c4b3a","Type":"ContainerDied","Data":"7eadb2058d2aa243856cd00ec267e67e2d158e64199a746e3460ff8c4b192b00"} Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.202323 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" podStartSLOduration=3.202301327 podStartE2EDuration="3.202301327s" podCreationTimestamp="2026-03-09 18:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:45:51.174770431 +0000 UTC m=+1232.517242839" watchObservedRunningTime="2026-03-09 18:45:51.202301327 +0000 UTC m=+1232.544773745" Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.249269 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fb8f999f-knd9k"] Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.279890 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66fb8f999f-knd9k"] Mar 09 18:45:51 crc kubenswrapper[4750]: I0309 18:45:51.399887 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a73090f-b06c-4eb9-bb78-6b91f20984d0" path="/var/lib/kubelet/pods/4a73090f-b06c-4eb9-bb78-6b91f20984d0/volumes" Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.167306 4750 generic.go:334] "Generic (PLEG): container finished" podID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerID="1624d9c60026bcd3e01d587842fe8b34482e04207adfcf064ea1215fd2120205" exitCode=0 Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.167410 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerDied","Data":"1624d9c60026bcd3e01d587842fe8b34482e04207adfcf064ea1215fd2120205"} Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.171520 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7fd7273f-010c-48a0-9ecc-bd80e0bff239","Type":"ContainerStarted","Data":"287eb011724f870e333b4e935d7583fdc4e6a2191fc36e0c77d3e572ab5c3399"} Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.173322 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"3926781f-b03f-45a7-82e5-10abf0c16c1e","Type":"ContainerStarted","Data":"ef16edbd1f978bdde45acc1dde08e21f76e7d8c166b9ac45794ad9861eaa07c0"} Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.176333 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6dgcc" event={"ID":"89841873-12eb-4add-bf8a-061b624c4b3a","Type":"ContainerStarted","Data":"f07d0b2b1864c406cebd9c79e3e77abc3cbf789813350f63505e9470a6f43977"} Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.176406 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6dgcc" event={"ID":"89841873-12eb-4add-bf8a-061b624c4b3a","Type":"ContainerStarted","Data":"3b5da458e5749ead9720e34edf460bb432e9f8701e11c908cfa92b0639181792"} Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.176687 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.176716 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:45:52 crc kubenswrapper[4750]: I0309 18:45:52.223198 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6dgcc" podStartSLOduration=10.071017016 podStartE2EDuration="34.2231823s" podCreationTimestamp="2026-03-09 18:45:18 +0000 UTC" firstStartedPulling="2026-03-09 18:45:26.312264762 +0000 UTC m=+1207.654737160" lastFinishedPulling="2026-03-09 18:45:50.464430046 +0000 UTC m=+1231.806902444" observedRunningTime="2026-03-09 18:45:52.220660721 +0000 UTC m=+1233.563133129" watchObservedRunningTime="2026-03-09 18:45:52.2231823 +0000 UTC m=+1233.565654698" Mar 09 18:45:54 crc kubenswrapper[4750]: I0309 18:45:54.194583 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n5qz8" event={"ID":"5e9fd826-3829-4379-9717-1b76ef376a39","Type":"ContainerStarted","Data":"1942d6ed9831efc2a8b159ba9710291e1ead063306d15710abec149ad39696ce"} Mar 09 18:45:54 crc kubenswrapper[4750]: I0309 18:45:54.195357 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-n5qz8" Mar 09 18:45:54 crc kubenswrapper[4750]: I0309 18:45:54.220357 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-n5qz8" podStartSLOduration=8.756267077 podStartE2EDuration="36.220330985s" podCreationTimestamp="2026-03-09 18:45:18 +0000 UTC" firstStartedPulling="2026-03-09 18:45:25.979528643 +0000 UTC m=+1207.322001041" lastFinishedPulling="2026-03-09 18:45:53.443592541 +0000 UTC m=+1234.786064949" observedRunningTime="2026-03-09 18:45:54.213361516 +0000 UTC m=+1235.555833924" watchObservedRunningTime="2026-03-09 18:45:54.220330985 +0000 UTC m=+1235.562803383" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.230723 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"64ff0811-8670-4605-9ef6-383f43259ace","Type":"ContainerStarted","Data":"b0968c2bd4cee8b2f094ec3aa22750800866cfe3a01f5dd80c9ea696958866c4"} Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.234050 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"302428aa-7eba-4f8b-8868-1e8e883d38c2","Type":"ContainerStarted","Data":"586a09c73313538aa721476b485136bd5ae237173823fb3c0aa4fdb854ceae02"} Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.606062 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cb4b9444f-7dhb2"] Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.606919 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerName="dnsmasq-dns" containerID="cri-o://c2e32c99687231ab22b644db43f701f17bb17861132261104abb1a5927b7a0a7" gracePeriod=10 Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.616382 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.639185 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cd9f6b887-2nlmb"] Mar 09 18:45:56 crc kubenswrapper[4750]: E0309 18:45:56.639751 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a73090f-b06c-4eb9-bb78-6b91f20984d0" containerName="init" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.639769 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a73090f-b06c-4eb9-bb78-6b91f20984d0" containerName="init" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.639942 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a73090f-b06c-4eb9-bb78-6b91f20984d0" containerName="init" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.642490 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.676695 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cd9f6b887-2nlmb"] Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.743852 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s42r\" (UniqueName: \"kubernetes.io/projected/66f23451-0dce-4bfa-95e7-8d17adc315d0-kube-api-access-2s42r\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.743931 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.743961 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-config\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.743981 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.744073 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-dns-svc\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.845985 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s42r\" (UniqueName: \"kubernetes.io/projected/66f23451-0dce-4bfa-95e7-8d17adc315d0-kube-api-access-2s42r\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.846058 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.846084 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-config\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.846099 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.846172 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-dns-svc\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.847225 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-dns-svc\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.848099 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-config\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.848375 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.848593 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.870864 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s42r\" (UniqueName: \"kubernetes.io/projected/66f23451-0dce-4bfa-95e7-8d17adc315d0-kube-api-access-2s42r\") pod \"dnsmasq-dns-7cd9f6b887-2nlmb\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:56 crc kubenswrapper[4750]: I0309 18:45:56.959863 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.544837 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cd9f6b887-2nlmb"] Mar 09 18:45:57 crc kubenswrapper[4750]: W0309 18:45:57.553058 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66f23451_0dce_4bfa_95e7_8d17adc315d0.slice/crio-b81a8455253a9b8b62c93750d5fcc0be564f9fd1b2dc7974773e0b02e31a8a2a WatchSource:0}: Error finding container b81a8455253a9b8b62c93750d5fcc0be564f9fd1b2dc7974773e0b02e31a8a2a: Status 404 returned error can't find the container with id b81a8455253a9b8b62c93750d5fcc0be564f9fd1b2dc7974773e0b02e31a8a2a Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.808479 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.814908 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.818556 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.818854 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-p8pg5" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.818948 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.819032 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.833660 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.979471 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cd12d247-7fd6-468d-8450-c395c4ee57ac-lock\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.979539 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.979616 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.979650 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cd12d247-7fd6-468d-8450-c395c4ee57ac-cache\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.979685 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnfj6\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-kube-api-access-bnfj6\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:57 crc kubenswrapper[4750]: I0309 18:45:57.979720 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd12d247-7fd6-468d-8450-c395c4ee57ac-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.081336 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.081386 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cd12d247-7fd6-468d-8450-c395c4ee57ac-cache\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.081431 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnfj6\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-kube-api-access-bnfj6\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.081470 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd12d247-7fd6-468d-8450-c395c4ee57ac-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.081489 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cd12d247-7fd6-468d-8450-c395c4ee57ac-lock\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.081517 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: E0309 18:45:58.081743 4750 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 09 18:45:58 crc kubenswrapper[4750]: E0309 18:45:58.081758 4750 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 09 18:45:58 crc kubenswrapper[4750]: E0309 18:45:58.081812 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift podName:cd12d247-7fd6-468d-8450-c395c4ee57ac nodeName:}" failed. No retries permitted until 2026-03-09 18:45:58.581791776 +0000 UTC m=+1239.924264174 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift") pod "swift-storage-0" (UID: "cd12d247-7fd6-468d-8450-c395c4ee57ac") : configmap "swift-ring-files" not found Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.081906 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.082237 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cd12d247-7fd6-468d-8450-c395c4ee57ac-cache\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.082293 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cd12d247-7fd6-468d-8450-c395c4ee57ac-lock\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.097889 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd12d247-7fd6-468d-8450-c395c4ee57ac-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.103308 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnfj6\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-kube-api-access-bnfj6\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.104545 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.261430 4750 generic.go:334] "Generic (PLEG): container finished" podID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerID="75ae4399a621829d965c2412052f8815980f3acb9857214f9823c0287999a261" exitCode=0 Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.261555 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" event={"ID":"66f23451-0dce-4bfa-95e7-8d17adc315d0","Type":"ContainerDied","Data":"75ae4399a621829d965c2412052f8815980f3acb9857214f9823c0287999a261"} Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.261931 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" event={"ID":"66f23451-0dce-4bfa-95e7-8d17adc315d0","Type":"ContainerStarted","Data":"b81a8455253a9b8b62c93750d5fcc0be564f9fd1b2dc7974773e0b02e31a8a2a"} Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.268415 4750 generic.go:334] "Generic (PLEG): container finished" podID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerID="c2e32c99687231ab22b644db43f701f17bb17861132261104abb1a5927b7a0a7" exitCode=0 Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.268504 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" event={"ID":"a90672b1-a38d-4f7a-b770-22fab99c5474","Type":"ContainerDied","Data":"c2e32c99687231ab22b644db43f701f17bb17861132261104abb1a5927b7a0a7"} Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.268875 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.325653 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.617346955 podStartE2EDuration="42.325615475s" podCreationTimestamp="2026-03-09 18:45:16 +0000 UTC" firstStartedPulling="2026-03-09 18:45:26.047256629 +0000 UTC m=+1207.389729017" lastFinishedPulling="2026-03-09 18:45:55.755525139 +0000 UTC m=+1237.097997537" observedRunningTime="2026-03-09 18:45:58.321472714 +0000 UTC m=+1239.663945112" watchObservedRunningTime="2026-03-09 18:45:58.325615475 +0000 UTC m=+1239.668087873" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.386614 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bjl5g"] Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.389415 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.393325 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.394194 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.394366 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.405838 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bjl5g"] Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.487777 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-swiftconf\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.487843 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-ring-data-devices\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.487905 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-combined-ca-bundle\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.488130 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-scripts\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.488171 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-dispersionconf\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.488200 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64b40944-803f-45ca-9d47-86c4250de34b-etc-swift\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.488229 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5qts\" (UniqueName: \"kubernetes.io/projected/64b40944-803f-45ca-9d47-86c4250de34b-kube-api-access-k5qts\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.589834 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-combined-ca-bundle\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.589904 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.589996 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-scripts\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.590028 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-dispersionconf\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.590051 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64b40944-803f-45ca-9d47-86c4250de34b-etc-swift\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.590077 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5qts\" (UniqueName: \"kubernetes.io/projected/64b40944-803f-45ca-9d47-86c4250de34b-kube-api-access-k5qts\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.590145 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-swiftconf\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.590183 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-ring-data-devices\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.590813 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-ring-data-devices\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.591030 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64b40944-803f-45ca-9d47-86c4250de34b-etc-swift\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: E0309 18:45:58.592361 4750 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 09 18:45:58 crc kubenswrapper[4750]: E0309 18:45:58.592392 4750 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 09 18:45:58 crc kubenswrapper[4750]: E0309 18:45:58.592438 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift podName:cd12d247-7fd6-468d-8450-c395c4ee57ac nodeName:}" failed. No retries permitted until 2026-03-09 18:45:59.592420978 +0000 UTC m=+1240.934893376 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift") pod "swift-storage-0" (UID: "cd12d247-7fd6-468d-8450-c395c4ee57ac") : configmap "swift-ring-files" not found Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.593112 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-scripts\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.595812 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-combined-ca-bundle\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.611112 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-swiftconf\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.617985 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-dispersionconf\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.628298 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5qts\" (UniqueName: \"kubernetes.io/projected/64b40944-803f-45ca-9d47-86c4250de34b-kube-api-access-k5qts\") pod \"swift-ring-rebalance-bjl5g\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:58 crc kubenswrapper[4750]: I0309 18:45:58.711668 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:45:59 crc kubenswrapper[4750]: I0309 18:45:59.611655 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:45:59 crc kubenswrapper[4750]: E0309 18:45:59.611902 4750 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 09 18:45:59 crc kubenswrapper[4750]: E0309 18:45:59.612138 4750 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 09 18:45:59 crc kubenswrapper[4750]: E0309 18:45:59.612203 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift podName:cd12d247-7fd6-468d-8450-c395c4ee57ac nodeName:}" failed. No retries permitted until 2026-03-09 18:46:01.61218613 +0000 UTC m=+1242.954658528 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift") pod "swift-storage-0" (UID: "cd12d247-7fd6-468d-8450-c395c4ee57ac") : configmap "swift-ring-files" not found Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.138792 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551366-ds24r"] Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.141108 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551366-ds24r" Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.144182 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.144945 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.145229 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.152435 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551366-ds24r"] Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.223938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtzlb\" (UniqueName: \"kubernetes.io/projected/7567fbed-2455-4d79-86fc-296893f49e2d-kube-api-access-wtzlb\") pod \"auto-csr-approver-29551366-ds24r\" (UID: \"7567fbed-2455-4d79-86fc-296893f49e2d\") " pod="openshift-infra/auto-csr-approver-29551366-ds24r" Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.326542 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtzlb\" (UniqueName: \"kubernetes.io/projected/7567fbed-2455-4d79-86fc-296893f49e2d-kube-api-access-wtzlb\") pod \"auto-csr-approver-29551366-ds24r\" (UID: \"7567fbed-2455-4d79-86fc-296893f49e2d\") " pod="openshift-infra/auto-csr-approver-29551366-ds24r" Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.352569 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtzlb\" (UniqueName: \"kubernetes.io/projected/7567fbed-2455-4d79-86fc-296893f49e2d-kube-api-access-wtzlb\") pod \"auto-csr-approver-29551366-ds24r\" (UID: \"7567fbed-2455-4d79-86fc-296893f49e2d\") " pod="openshift-infra/auto-csr-approver-29551366-ds24r" Mar 09 18:46:00 crc kubenswrapper[4750]: I0309 18:46:00.481478 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551366-ds24r" Mar 09 18:46:01 crc kubenswrapper[4750]: I0309 18:46:01.380285 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 18:46:01 crc kubenswrapper[4750]: I0309 18:46:01.651621 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:46:01 crc kubenswrapper[4750]: E0309 18:46:01.652057 4750 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 09 18:46:01 crc kubenswrapper[4750]: E0309 18:46:01.652118 4750 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 09 18:46:01 crc kubenswrapper[4750]: E0309 18:46:01.652221 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift podName:cd12d247-7fd6-468d-8450-c395c4ee57ac nodeName:}" failed. No retries permitted until 2026-03-09 18:46:05.652191698 +0000 UTC m=+1246.994664136 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift") pod "swift-storage-0" (UID: "cd12d247-7fd6-468d-8450-c395c4ee57ac") : configmap "swift-ring-files" not found Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.047668 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.159592 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-sb\") pod \"a90672b1-a38d-4f7a-b770-22fab99c5474\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.160024 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztb4q\" (UniqueName: \"kubernetes.io/projected/a90672b1-a38d-4f7a-b770-22fab99c5474-kube-api-access-ztb4q\") pod \"a90672b1-a38d-4f7a-b770-22fab99c5474\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.160101 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-dns-svc\") pod \"a90672b1-a38d-4f7a-b770-22fab99c5474\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.160142 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-nb\") pod \"a90672b1-a38d-4f7a-b770-22fab99c5474\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.160184 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-config\") pod \"a90672b1-a38d-4f7a-b770-22fab99c5474\" (UID: \"a90672b1-a38d-4f7a-b770-22fab99c5474\") " Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.175122 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90672b1-a38d-4f7a-b770-22fab99c5474-kube-api-access-ztb4q" (OuterVolumeSpecName: "kube-api-access-ztb4q") pod "a90672b1-a38d-4f7a-b770-22fab99c5474" (UID: "a90672b1-a38d-4f7a-b770-22fab99c5474"). InnerVolumeSpecName "kube-api-access-ztb4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.238793 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a90672b1-a38d-4f7a-b770-22fab99c5474" (UID: "a90672b1-a38d-4f7a-b770-22fab99c5474"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.249169 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a90672b1-a38d-4f7a-b770-22fab99c5474" (UID: "a90672b1-a38d-4f7a-b770-22fab99c5474"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.249362 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-config" (OuterVolumeSpecName: "config") pod "a90672b1-a38d-4f7a-b770-22fab99c5474" (UID: "a90672b1-a38d-4f7a-b770-22fab99c5474"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.258067 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a90672b1-a38d-4f7a-b770-22fab99c5474" (UID: "a90672b1-a38d-4f7a-b770-22fab99c5474"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.266089 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztb4q\" (UniqueName: \"kubernetes.io/projected/a90672b1-a38d-4f7a-b770-22fab99c5474-kube-api-access-ztb4q\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.266119 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.266130 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.266138 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.266146 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a90672b1-a38d-4f7a-b770-22fab99c5474-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.309045 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" event={"ID":"a90672b1-a38d-4f7a-b770-22fab99c5474","Type":"ContainerDied","Data":"f0e382f70d3048534c32f268cb080a800b0dbe622de0aca60bb89138b6f6f8c4"} Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.309118 4750 scope.go:117] "RemoveContainer" containerID="c2e32c99687231ab22b644db43f701f17bb17861132261104abb1a5927b7a0a7" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.309157 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.371227 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cb4b9444f-7dhb2"] Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.371296 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cb4b9444f-7dhb2"] Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.371468 4750 scope.go:117] "RemoveContainer" containerID="d768357ba96d4a2d8d721908bf89b7b1539d671d06c5c42d87f18482da335dcf" Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.600149 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551366-ds24r"] Mar 09 18:46:02 crc kubenswrapper[4750]: W0309 18:46:02.603506 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7567fbed_2455_4d79_86fc_296893f49e2d.slice/crio-99630e1dcc637f7e1f778dad0888b14a013c24ca4b8f229ff686df723295c59e WatchSource:0}: Error finding container 99630e1dcc637f7e1f778dad0888b14a013c24ca4b8f229ff686df723295c59e: Status 404 returned error can't find the container with id 99630e1dcc637f7e1f778dad0888b14a013c24ca4b8f229ff686df723295c59e Mar 09 18:46:02 crc kubenswrapper[4750]: W0309 18:46:02.644424 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64b40944_803f_45ca_9d47_86c4250de34b.slice/crio-6ffb7f5243953fac065a217f67e3e7c35f664073d9e3157c7d50ca56b5d47061 WatchSource:0}: Error finding container 6ffb7f5243953fac065a217f67e3e7c35f664073d9e3157c7d50ca56b5d47061: Status 404 returned error can't find the container with id 6ffb7f5243953fac065a217f67e3e7c35f664073d9e3157c7d50ca56b5d47061 Mar 09 18:46:02 crc kubenswrapper[4750]: I0309 18:46:02.644693 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bjl5g"] Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.320993 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f059f39a-1c7f-46af-be90-4095105df733","Type":"ContainerStarted","Data":"264664e9748832f13eb56ad2816d80b6ef053f7466f2f7fd9c23500578346e44"} Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.325129 4750 generic.go:334] "Generic (PLEG): container finished" podID="7fd7273f-010c-48a0-9ecc-bd80e0bff239" containerID="287eb011724f870e333b4e935d7583fdc4e6a2191fc36e0c77d3e572ab5c3399" exitCode=0 Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.325188 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7fd7273f-010c-48a0-9ecc-bd80e0bff239","Type":"ContainerDied","Data":"287eb011724f870e333b4e935d7583fdc4e6a2191fc36e0c77d3e572ab5c3399"} Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.328121 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" event={"ID":"66f23451-0dce-4bfa-95e7-8d17adc315d0","Type":"ContainerStarted","Data":"49fbad12eb84b0f3d14f848f220cc82823fd7871d7abf0b485264235d15dc73f"} Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.328250 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.329682 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bjl5g" event={"ID":"64b40944-803f-45ca-9d47-86c4250de34b","Type":"ContainerStarted","Data":"6ffb7f5243953fac065a217f67e3e7c35f664073d9e3157c7d50ca56b5d47061"} Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.331528 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerStarted","Data":"fdc04f5cb47fe3f3220d52e717ae63c34c4ac331da7289cd7951365dbf2bfdde"} Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.332342 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551366-ds24r" event={"ID":"7567fbed-2455-4d79-86fc-296893f49e2d","Type":"ContainerStarted","Data":"99630e1dcc637f7e1f778dad0888b14a013c24ca4b8f229ff686df723295c59e"} Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.345326 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.456718855 podStartE2EDuration="44.345307053s" podCreationTimestamp="2026-03-09 18:45:19 +0000 UTC" firstStartedPulling="2026-03-09 18:45:26.275309871 +0000 UTC m=+1207.617782269" lastFinishedPulling="2026-03-09 18:46:02.163898049 +0000 UTC m=+1243.506370467" observedRunningTime="2026-03-09 18:46:03.340758049 +0000 UTC m=+1244.683230457" watchObservedRunningTime="2026-03-09 18:46:03.345307053 +0000 UTC m=+1244.687779451" Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.387405 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" path="/var/lib/kubelet/pods/a90672b1-a38d-4f7a-b770-22fab99c5474/volumes" Mar 09 18:46:03 crc kubenswrapper[4750]: I0309 18:46:03.389530 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" podStartSLOduration=7.38948911 podStartE2EDuration="7.38948911s" podCreationTimestamp="2026-03-09 18:45:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:03.387040003 +0000 UTC m=+1244.729512401" watchObservedRunningTime="2026-03-09 18:46:03.38948911 +0000 UTC m=+1244.731961518" Mar 09 18:46:04 crc kubenswrapper[4750]: I0309 18:46:04.236996 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cb4b9444f-7dhb2" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: i/o timeout" Mar 09 18:46:05 crc kubenswrapper[4750]: I0309 18:46:05.356130 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerStarted","Data":"aa74b7401b76e1275fb507277ec22d2dff67430bfb85f3824ad625b656f40b80"} Mar 09 18:46:05 crc kubenswrapper[4750]: I0309 18:46:05.737073 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:46:05 crc kubenswrapper[4750]: E0309 18:46:05.737250 4750 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 09 18:46:05 crc kubenswrapper[4750]: E0309 18:46:05.737367 4750 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 09 18:46:05 crc kubenswrapper[4750]: E0309 18:46:05.737421 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift podName:cd12d247-7fd6-468d-8450-c395c4ee57ac nodeName:}" failed. No retries permitted until 2026-03-09 18:46:13.737405664 +0000 UTC m=+1255.079878062 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift") pod "swift-storage-0" (UID: "cd12d247-7fd6-468d-8450-c395c4ee57ac") : configmap "swift-ring-files" not found Mar 09 18:46:05 crc kubenswrapper[4750]: I0309 18:46:05.887429 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 09 18:46:05 crc kubenswrapper[4750]: I0309 18:46:05.887494 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 09 18:46:05 crc kubenswrapper[4750]: I0309 18:46:05.941547 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.369740 4750 generic.go:334] "Generic (PLEG): container finished" podID="302428aa-7eba-4f8b-8868-1e8e883d38c2" containerID="586a09c73313538aa721476b485136bd5ae237173823fb3c0aa4fdb854ceae02" exitCode=0 Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.369827 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"302428aa-7eba-4f8b-8868-1e8e883d38c2","Type":"ContainerDied","Data":"586a09c73313538aa721476b485136bd5ae237173823fb3c0aa4fdb854ceae02"} Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.378900 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bjl5g" event={"ID":"64b40944-803f-45ca-9d47-86c4250de34b","Type":"ContainerStarted","Data":"8a5190041440d112791d0a695e58b5491459eed28292c0abd9dba0b0473b23bf"} Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.382899 4750 generic.go:334] "Generic (PLEG): container finished" podID="7567fbed-2455-4d79-86fc-296893f49e2d" containerID="f719e033a147462c6c3db9db045bd1371c2b1a1f7a8cf0b4284eae84a2d5ecc2" exitCode=0 Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.382950 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551366-ds24r" event={"ID":"7567fbed-2455-4d79-86fc-296893f49e2d","Type":"ContainerDied","Data":"f719e033a147462c6c3db9db045bd1371c2b1a1f7a8cf0b4284eae84a2d5ecc2"} Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.386410 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7fd7273f-010c-48a0-9ecc-bd80e0bff239","Type":"ContainerStarted","Data":"1a739ae5c30b684a563c106b5b9664ea9deb17b55708c62f3937a48a973ce6d9"} Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.453442 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bjl5g" podStartSLOduration=5.819907087 podStartE2EDuration="8.453419642s" podCreationTimestamp="2026-03-09 18:45:58 +0000 UTC" firstStartedPulling="2026-03-09 18:46:02.653440618 +0000 UTC m=+1243.995913016" lastFinishedPulling="2026-03-09 18:46:05.286953163 +0000 UTC m=+1246.629425571" observedRunningTime="2026-03-09 18:46:06.435545888 +0000 UTC m=+1247.778018306" watchObservedRunningTime="2026-03-09 18:46:06.453419642 +0000 UTC m=+1247.795892040" Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.481689 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=29.303785631 podStartE2EDuration="54.481665019s" podCreationTimestamp="2026-03-09 18:45:12 +0000 UTC" firstStartedPulling="2026-03-09 18:45:26.27604478 +0000 UTC m=+1207.618517178" lastFinishedPulling="2026-03-09 18:45:51.453924158 +0000 UTC m=+1232.796396566" observedRunningTime="2026-03-09 18:46:06.476246792 +0000 UTC m=+1247.818719220" watchObservedRunningTime="2026-03-09 18:46:06.481665019 +0000 UTC m=+1247.824137417" Mar 09 18:46:06 crc kubenswrapper[4750]: I0309 18:46:06.502183 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 09 18:46:07 crc kubenswrapper[4750]: I0309 18:46:07.397990 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"302428aa-7eba-4f8b-8868-1e8e883d38c2","Type":"ContainerStarted","Data":"153a4d54fdbed31db60733eaa23607a5bd545c1f22f6d9a5d5133ab57f59da7e"} Mar 09 18:46:09 crc kubenswrapper[4750]: I0309 18:46:09.571599 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551366-ds24r" Mar 09 18:46:09 crc kubenswrapper[4750]: I0309 18:46:09.594100 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371977.260704 podStartE2EDuration="59.594071435s" podCreationTimestamp="2026-03-09 18:45:10 +0000 UTC" firstStartedPulling="2026-03-09 18:45:26.280060729 +0000 UTC m=+1207.622533117" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:07.424790203 +0000 UTC m=+1248.767262601" watchObservedRunningTime="2026-03-09 18:46:09.594071435 +0000 UTC m=+1250.936543843" Mar 09 18:46:09 crc kubenswrapper[4750]: I0309 18:46:09.664923 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtzlb\" (UniqueName: \"kubernetes.io/projected/7567fbed-2455-4d79-86fc-296893f49e2d-kube-api-access-wtzlb\") pod \"7567fbed-2455-4d79-86fc-296893f49e2d\" (UID: \"7567fbed-2455-4d79-86fc-296893f49e2d\") " Mar 09 18:46:09 crc kubenswrapper[4750]: I0309 18:46:09.674851 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7567fbed-2455-4d79-86fc-296893f49e2d-kube-api-access-wtzlb" (OuterVolumeSpecName: "kube-api-access-wtzlb") pod "7567fbed-2455-4d79-86fc-296893f49e2d" (UID: "7567fbed-2455-4d79-86fc-296893f49e2d"). InnerVolumeSpecName "kube-api-access-wtzlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:09 crc kubenswrapper[4750]: I0309 18:46:09.769065 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtzlb\" (UniqueName: \"kubernetes.io/projected/7567fbed-2455-4d79-86fc-296893f49e2d-kube-api-access-wtzlb\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.446306 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerStarted","Data":"9e6e56cc11a17a86f0ea07eeea6eaa87b9ce12622f50dc0e6e792d2f5213dff5"} Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.451234 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551366-ds24r" event={"ID":"7567fbed-2455-4d79-86fc-296893f49e2d","Type":"ContainerDied","Data":"99630e1dcc637f7e1f778dad0888b14a013c24ca4b8f229ff686df723295c59e"} Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.451285 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99630e1dcc637f7e1f778dad0888b14a013c24ca4b8f229ff686df723295c59e" Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.451335 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551366-ds24r" Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.492303 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=10.964953416 podStartE2EDuration="54.492268833s" podCreationTimestamp="2026-03-09 18:45:16 +0000 UTC" firstStartedPulling="2026-03-09 18:45:26.063583722 +0000 UTC m=+1207.406056120" lastFinishedPulling="2026-03-09 18:46:09.590899139 +0000 UTC m=+1250.933371537" observedRunningTime="2026-03-09 18:46:10.484660507 +0000 UTC m=+1251.827132905" watchObservedRunningTime="2026-03-09 18:46:10.492268833 +0000 UTC m=+1251.834741231" Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.664318 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551360-plcvg"] Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.676174 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551360-plcvg"] Mar 09 18:46:10 crc kubenswrapper[4750]: I0309 18:46:10.928750 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.116277 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 09 18:46:11 crc kubenswrapper[4750]: E0309 18:46:11.116796 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7567fbed-2455-4d79-86fc-296893f49e2d" containerName="oc" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.116816 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7567fbed-2455-4d79-86fc-296893f49e2d" containerName="oc" Mar 09 18:46:11 crc kubenswrapper[4750]: E0309 18:46:11.116829 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerName="dnsmasq-dns" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.116836 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerName="dnsmasq-dns" Mar 09 18:46:11 crc kubenswrapper[4750]: E0309 18:46:11.116850 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerName="init" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.116857 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerName="init" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.117028 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90672b1-a38d-4f7a-b770-22fab99c5474" containerName="dnsmasq-dns" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.117048 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7567fbed-2455-4d79-86fc-296893f49e2d" containerName="oc" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.118133 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.120743 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.120981 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7jbxl" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.121604 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.121738 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.189832 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.195905 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-config\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.197125 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.197336 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.197669 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.198175 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-scripts\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.198492 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.199189 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdlt6\" (UniqueName: \"kubernetes.io/projected/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-kube-api-access-xdlt6\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.302737 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-scripts\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.303167 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.303359 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdlt6\" (UniqueName: \"kubernetes.io/projected/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-kube-api-access-xdlt6\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.303981 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-config\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.304002 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.304318 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.304479 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.304685 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.304980 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-config\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.305492 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-scripts\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.309390 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.324394 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.328158 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdlt6\" (UniqueName: \"kubernetes.io/projected/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-kube-api-access-xdlt6\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.329960 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fe00ea1-72f0-4ba3-8096-11e8ec17c733-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2fe00ea1-72f0-4ba3-8096-11e8ec17c733\") " pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.383320 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63" path="/var/lib/kubelet/pods/5c38630e-d1c0-4b5a-8c71-0e3b2f5d3f63/volumes" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.497955 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 09 18:46:11 crc kubenswrapper[4750]: I0309 18:46:11.962899 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.007462 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.058315 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-595d94d48f-57ntm"] Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.058673 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" podUID="e8ab0211-a492-45f9-9606-08706461761d" containerName="dnsmasq-dns" containerID="cri-o://e2a62efbca093ecd19a98c3d88d2dbb4b29f57dbe52ac80b0bc465b40db13bd6" gracePeriod=10 Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.366239 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.366289 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.519978 4750 generic.go:334] "Generic (PLEG): container finished" podID="e8ab0211-a492-45f9-9606-08706461761d" containerID="e2a62efbca093ecd19a98c3d88d2dbb4b29f57dbe52ac80b0bc465b40db13bd6" exitCode=0 Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.520095 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" event={"ID":"e8ab0211-a492-45f9-9606-08706461761d","Type":"ContainerDied","Data":"e2a62efbca093ecd19a98c3d88d2dbb4b29f57dbe52ac80b0bc465b40db13bd6"} Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.524737 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2fe00ea1-72f0-4ba3-8096-11e8ec17c733","Type":"ContainerStarted","Data":"921e806c1d5b0d475d50e888b60bfab56be48c55b09cfe790d16dfe36885fdbb"} Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.772765 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.823515 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.937350 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68fsb\" (UniqueName: \"kubernetes.io/projected/e8ab0211-a492-45f9-9606-08706461761d-kube-api-access-68fsb\") pod \"e8ab0211-a492-45f9-9606-08706461761d\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.937455 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-config\") pod \"e8ab0211-a492-45f9-9606-08706461761d\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.937559 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-dns-svc\") pod \"e8ab0211-a492-45f9-9606-08706461761d\" (UID: \"e8ab0211-a492-45f9-9606-08706461761d\") " Mar 09 18:46:12 crc kubenswrapper[4750]: I0309 18:46:12.943960 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ab0211-a492-45f9-9606-08706461761d-kube-api-access-68fsb" (OuterVolumeSpecName: "kube-api-access-68fsb") pod "e8ab0211-a492-45f9-9606-08706461761d" (UID: "e8ab0211-a492-45f9-9606-08706461761d"). InnerVolumeSpecName "kube-api-access-68fsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.000996 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-config" (OuterVolumeSpecName: "config") pod "e8ab0211-a492-45f9-9606-08706461761d" (UID: "e8ab0211-a492-45f9-9606-08706461761d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.002817 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8ab0211-a492-45f9-9606-08706461761d" (UID: "e8ab0211-a492-45f9-9606-08706461761d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.041472 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68fsb\" (UniqueName: \"kubernetes.io/projected/e8ab0211-a492-45f9-9606-08706461761d-kube-api-access-68fsb\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.041520 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.041531 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8ab0211-a492-45f9-9606-08706461761d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.537113 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" event={"ID":"e8ab0211-a492-45f9-9606-08706461761d","Type":"ContainerDied","Data":"abb1579db2a267921f130faefa34e53728a4bf58b06726024c7ed1df19a774b6"} Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.537426 4750 scope.go:117] "RemoveContainer" containerID="e2a62efbca093ecd19a98c3d88d2dbb4b29f57dbe52ac80b0bc465b40db13bd6" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.537542 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595d94d48f-57ntm" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.540118 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2fe00ea1-72f0-4ba3-8096-11e8ec17c733","Type":"ContainerStarted","Data":"96edb38d0e7b20255269bbbfeae24628d3ddb2ee3967357888da9a7961da6b8d"} Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.567058 4750 scope.go:117] "RemoveContainer" containerID="cc0b2db5e27692056731cd281a566068c49541e0931b36c5c30539f34b00a05a" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.573973 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-595d94d48f-57ntm"] Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.582527 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-595d94d48f-57ntm"] Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.753839 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:46:13 crc kubenswrapper[4750]: E0309 18:46:13.754108 4750 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 09 18:46:13 crc kubenswrapper[4750]: E0309 18:46:13.754132 4750 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 09 18:46:13 crc kubenswrapper[4750]: E0309 18:46:13.754190 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift podName:cd12d247-7fd6-468d-8450-c395c4ee57ac nodeName:}" failed. No retries permitted until 2026-03-09 18:46:29.754172101 +0000 UTC m=+1271.096644509 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift") pod "swift-storage-0" (UID: "cd12d247-7fd6-468d-8450-c395c4ee57ac") : configmap "swift-ring-files" not found Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.868798 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 09 18:46:13 crc kubenswrapper[4750]: I0309 18:46:13.869474 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.036272 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.563131 4750 generic.go:334] "Generic (PLEG): container finished" podID="64b40944-803f-45ca-9d47-86c4250de34b" containerID="8a5190041440d112791d0a695e58b5491459eed28292c0abd9dba0b0473b23bf" exitCode=0 Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.563235 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bjl5g" event={"ID":"64b40944-803f-45ca-9d47-86c4250de34b","Type":"ContainerDied","Data":"8a5190041440d112791d0a695e58b5491459eed28292c0abd9dba0b0473b23bf"} Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.569451 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2fe00ea1-72f0-4ba3-8096-11e8ec17c733","Type":"ContainerStarted","Data":"f4f698e3e192ba97ba1e2fd67ea21014080cfba2dd0e2ef0a575efb4bf481ee9"} Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.569502 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.640194 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.450068069 podStartE2EDuration="3.640160918s" podCreationTimestamp="2026-03-09 18:46:11 +0000 UTC" firstStartedPulling="2026-03-09 18:46:12.017033104 +0000 UTC m=+1253.359505502" lastFinishedPulling="2026-03-09 18:46:13.207125953 +0000 UTC m=+1254.549598351" observedRunningTime="2026-03-09 18:46:14.619749634 +0000 UTC m=+1255.962222052" watchObservedRunningTime="2026-03-09 18:46:14.640160918 +0000 UTC m=+1255.982633316" Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.710290 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 09 18:46:14 crc kubenswrapper[4750]: I0309 18:46:14.996511 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.141171 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.385471 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ab0211-a492-45f9-9606-08706461761d" path="/var/lib/kubelet/pods/e8ab0211-a492-45f9-9606-08706461761d/volumes" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.430299 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1af5-account-create-update-c74fv"] Mar 09 18:46:15 crc kubenswrapper[4750]: E0309 18:46:15.430705 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ab0211-a492-45f9-9606-08706461761d" containerName="dnsmasq-dns" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.430728 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ab0211-a492-45f9-9606-08706461761d" containerName="dnsmasq-dns" Mar 09 18:46:15 crc kubenswrapper[4750]: E0309 18:46:15.430757 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ab0211-a492-45f9-9606-08706461761d" containerName="init" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.430765 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ab0211-a492-45f9-9606-08706461761d" containerName="init" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.430935 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ab0211-a492-45f9-9606-08706461761d" containerName="dnsmasq-dns" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.431501 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.436750 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.445585 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1af5-account-create-update-c74fv"] Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.497666 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/567fa16f-93f1-430f-a041-aba3aaec9957-operator-scripts\") pod \"placement-1af5-account-create-update-c74fv\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.497719 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqssj\" (UniqueName: \"kubernetes.io/projected/567fa16f-93f1-430f-a041-aba3aaec9957-kube-api-access-hqssj\") pod \"placement-1af5-account-create-update-c74fv\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.499573 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-p5r85"] Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.500662 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p5r85" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.546268 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-p5r85"] Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.594548 4750 generic.go:334] "Generic (PLEG): container finished" podID="378d9b58-f830-4d61-b408-26668a301507" containerID="bbc618a451438e977e10a76341e34bb8d2c1f770d1a8c8ffc94ff70b158dfb10" exitCode=0 Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.594774 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"378d9b58-f830-4d61-b408-26668a301507","Type":"ContainerDied","Data":"bbc618a451438e977e10a76341e34bb8d2c1f770d1a8c8ffc94ff70b158dfb10"} Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.598910 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bplsz\" (UniqueName: \"kubernetes.io/projected/5bf233bd-79f7-4a07-8328-e7491ebb967b-kube-api-access-bplsz\") pod \"placement-db-create-p5r85\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " pod="openstack/placement-db-create-p5r85" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.598970 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf233bd-79f7-4a07-8328-e7491ebb967b-operator-scripts\") pod \"placement-db-create-p5r85\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " pod="openstack/placement-db-create-p5r85" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.599065 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/567fa16f-93f1-430f-a041-aba3aaec9957-operator-scripts\") pod \"placement-1af5-account-create-update-c74fv\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.599092 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqssj\" (UniqueName: \"kubernetes.io/projected/567fa16f-93f1-430f-a041-aba3aaec9957-kube-api-access-hqssj\") pod \"placement-1af5-account-create-update-c74fv\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.600084 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/567fa16f-93f1-430f-a041-aba3aaec9957-operator-scripts\") pod \"placement-1af5-account-create-update-c74fv\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.632840 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqssj\" (UniqueName: \"kubernetes.io/projected/567fa16f-93f1-430f-a041-aba3aaec9957-kube-api-access-hqssj\") pod \"placement-1af5-account-create-update-c74fv\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.700983 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bplsz\" (UniqueName: \"kubernetes.io/projected/5bf233bd-79f7-4a07-8328-e7491ebb967b-kube-api-access-bplsz\") pod \"placement-db-create-p5r85\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " pod="openstack/placement-db-create-p5r85" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.701121 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf233bd-79f7-4a07-8328-e7491ebb967b-operator-scripts\") pod \"placement-db-create-p5r85\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " pod="openstack/placement-db-create-p5r85" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.703958 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf233bd-79f7-4a07-8328-e7491ebb967b-operator-scripts\") pod \"placement-db-create-p5r85\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " pod="openstack/placement-db-create-p5r85" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.747143 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.757687 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bplsz\" (UniqueName: \"kubernetes.io/projected/5bf233bd-79f7-4a07-8328-e7491ebb967b-kube-api-access-bplsz\") pod \"placement-db-create-p5r85\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " pod="openstack/placement-db-create-p5r85" Mar 09 18:46:15 crc kubenswrapper[4750]: I0309 18:46:15.831936 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p5r85" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.011356 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.108140 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-scripts\") pod \"64b40944-803f-45ca-9d47-86c4250de34b\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.108226 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-ring-data-devices\") pod \"64b40944-803f-45ca-9d47-86c4250de34b\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.108275 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-combined-ca-bundle\") pod \"64b40944-803f-45ca-9d47-86c4250de34b\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.108375 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-dispersionconf\") pod \"64b40944-803f-45ca-9d47-86c4250de34b\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.108458 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64b40944-803f-45ca-9d47-86c4250de34b-etc-swift\") pod \"64b40944-803f-45ca-9d47-86c4250de34b\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.108482 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-swiftconf\") pod \"64b40944-803f-45ca-9d47-86c4250de34b\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.108509 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5qts\" (UniqueName: \"kubernetes.io/projected/64b40944-803f-45ca-9d47-86c4250de34b-kube-api-access-k5qts\") pod \"64b40944-803f-45ca-9d47-86c4250de34b\" (UID: \"64b40944-803f-45ca-9d47-86c4250de34b\") " Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.110353 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "64b40944-803f-45ca-9d47-86c4250de34b" (UID: "64b40944-803f-45ca-9d47-86c4250de34b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.111762 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64b40944-803f-45ca-9d47-86c4250de34b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "64b40944-803f-45ca-9d47-86c4250de34b" (UID: "64b40944-803f-45ca-9d47-86c4250de34b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.116147 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64b40944-803f-45ca-9d47-86c4250de34b-kube-api-access-k5qts" (OuterVolumeSpecName: "kube-api-access-k5qts") pod "64b40944-803f-45ca-9d47-86c4250de34b" (UID: "64b40944-803f-45ca-9d47-86c4250de34b"). InnerVolumeSpecName "kube-api-access-k5qts". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.117965 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "64b40944-803f-45ca-9d47-86c4250de34b" (UID: "64b40944-803f-45ca-9d47-86c4250de34b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.137840 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64b40944-803f-45ca-9d47-86c4250de34b" (UID: "64b40944-803f-45ca-9d47-86c4250de34b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.141657 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-scripts" (OuterVolumeSpecName: "scripts") pod "64b40944-803f-45ca-9d47-86c4250de34b" (UID: "64b40944-803f-45ca-9d47-86c4250de34b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.142928 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "64b40944-803f-45ca-9d47-86c4250de34b" (UID: "64b40944-803f-45ca-9d47-86c4250de34b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.211053 4750 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.211100 4750 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64b40944-803f-45ca-9d47-86c4250de34b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.211110 4750 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.211121 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5qts\" (UniqueName: \"kubernetes.io/projected/64b40944-803f-45ca-9d47-86c4250de34b-kube-api-access-k5qts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.211136 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.211147 4750 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64b40944-803f-45ca-9d47-86c4250de34b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.211157 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b40944-803f-45ca-9d47-86c4250de34b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.307882 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1af5-account-create-update-c74fv"] Mar 09 18:46:16 crc kubenswrapper[4750]: W0309 18:46:16.310399 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod567fa16f_93f1_430f_a041_aba3aaec9957.slice/crio-be1b7db7c1b4626d0e4e108a3637126126750d4cb941fd2e409605c4418e01fe WatchSource:0}: Error finding container be1b7db7c1b4626d0e4e108a3637126126750d4cb941fd2e409605c4418e01fe: Status 404 returned error can't find the container with id be1b7db7c1b4626d0e4e108a3637126126750d4cb941fd2e409605c4418e01fe Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.501039 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-p5r85"] Mar 09 18:46:16 crc kubenswrapper[4750]: W0309 18:46:16.510377 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bf233bd_79f7_4a07_8328_e7491ebb967b.slice/crio-a78485dfa29d62f8c3ba528b4eb2cce15aaa1eafd9d0aa5b22f070d3f80225bd WatchSource:0}: Error finding container a78485dfa29d62f8c3ba528b4eb2cce15aaa1eafd9d0aa5b22f070d3f80225bd: Status 404 returned error can't find the container with id a78485dfa29d62f8c3ba528b4eb2cce15aaa1eafd9d0aa5b22f070d3f80225bd Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.591145 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-f9d9-account-create-update-868dt"] Mar 09 18:46:16 crc kubenswrapper[4750]: E0309 18:46:16.591745 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b40944-803f-45ca-9d47-86c4250de34b" containerName="swift-ring-rebalance" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.591773 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b40944-803f-45ca-9d47-86c4250de34b" containerName="swift-ring-rebalance" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.592014 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="64b40944-803f-45ca-9d47-86c4250de34b" containerName="swift-ring-rebalance" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.592984 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.595874 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.605406 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-qlj7j"] Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.609581 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.609772 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p5r85" event={"ID":"5bf233bd-79f7-4a07-8328-e7491ebb967b","Type":"ContainerStarted","Data":"a78485dfa29d62f8c3ba528b4eb2cce15aaa1eafd9d0aa5b22f070d3f80225bd"} Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.612316 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1af5-account-create-update-c74fv" event={"ID":"567fa16f-93f1-430f-a041-aba3aaec9957","Type":"ContainerStarted","Data":"af8da8c525aa15597d26252e499384ac9131562117fee699016088d024d65f34"} Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.612362 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1af5-account-create-update-c74fv" event={"ID":"567fa16f-93f1-430f-a041-aba3aaec9957","Type":"ContainerStarted","Data":"be1b7db7c1b4626d0e4e108a3637126126750d4cb941fd2e409605c4418e01fe"} Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.619550 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bjl5g" event={"ID":"64b40944-803f-45ca-9d47-86c4250de34b","Type":"ContainerDied","Data":"6ffb7f5243953fac065a217f67e3e7c35f664073d9e3157c7d50ca56b5d47061"} Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.619611 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ffb7f5243953fac065a217f67e3e7c35f664073d9e3157c7d50ca56b5d47061" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.619929 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bjl5g" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.654213 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-f9d9-account-create-update-868dt"] Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.654258 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"378d9b58-f830-4d61-b408-26668a301507","Type":"ContainerStarted","Data":"3ffaf36e1a2a216af73d4039972485467b465dcd9d2950724768a1894bcb9519"} Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.655168 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.658655 4750 generic.go:334] "Generic (PLEG): container finished" podID="3236a6fb-c288-441c-96e5-7941f818b0af" containerID="4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143" exitCode=0 Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.659744 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3236a6fb-c288-441c-96e5-7941f818b0af","Type":"ContainerDied","Data":"4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143"} Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.668793 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-qlj7j"] Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.717555 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c0825fd-c53a-46cc-9b8a-fca000620d76-operator-scripts\") pod \"watcher-f9d9-account-create-update-868dt\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.717638 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-597nb\" (UniqueName: \"kubernetes.io/projected/1c0825fd-c53a-46cc-9b8a-fca000620d76-kube-api-access-597nb\") pod \"watcher-f9d9-account-create-update-868dt\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.729302 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-1af5-account-create-update-c74fv" podStartSLOduration=1.729274846 podStartE2EDuration="1.729274846s" podCreationTimestamp="2026-03-09 18:46:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:16.638534066 +0000 UTC m=+1257.981006464" watchObservedRunningTime="2026-03-09 18:46:16.729274846 +0000 UTC m=+1258.071747244" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.784113 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.31951437 podStartE2EDuration="1m7.784089632s" podCreationTimestamp="2026-03-09 18:45:09 +0000 UTC" firstStartedPulling="2026-03-09 18:45:25.539216728 +0000 UTC m=+1206.881689126" lastFinishedPulling="2026-03-09 18:45:41.00379198 +0000 UTC m=+1222.346264388" observedRunningTime="2026-03-09 18:46:16.697458294 +0000 UTC m=+1258.039930692" watchObservedRunningTime="2026-03-09 18:46:16.784089632 +0000 UTC m=+1258.126562030" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.852404 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c0825fd-c53a-46cc-9b8a-fca000620d76-operator-scripts\") pod \"watcher-f9d9-account-create-update-868dt\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.853341 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-597nb\" (UniqueName: \"kubernetes.io/projected/1c0825fd-c53a-46cc-9b8a-fca000620d76-kube-api-access-597nb\") pod \"watcher-f9d9-account-create-update-868dt\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.853920 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbp79\" (UniqueName: \"kubernetes.io/projected/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-kube-api-access-mbp79\") pod \"watcher-db-create-qlj7j\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.854071 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-operator-scripts\") pod \"watcher-db-create-qlj7j\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.856996 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c0825fd-c53a-46cc-9b8a-fca000620d76-operator-scripts\") pod \"watcher-f9d9-account-create-update-868dt\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.889586 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-597nb\" (UniqueName: \"kubernetes.io/projected/1c0825fd-c53a-46cc-9b8a-fca000620d76-kube-api-access-597nb\") pod \"watcher-f9d9-account-create-update-868dt\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.891232 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.960456 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbp79\" (UniqueName: \"kubernetes.io/projected/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-kube-api-access-mbp79\") pod \"watcher-db-create-qlj7j\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.960539 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-operator-scripts\") pod \"watcher-db-create-qlj7j\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.961900 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-operator-scripts\") pod \"watcher-db-create-qlj7j\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:16 crc kubenswrapper[4750]: I0309 18:46:16.979899 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbp79\" (UniqueName: \"kubernetes.io/projected/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-kube-api-access-mbp79\") pod \"watcher-db-create-qlj7j\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.191055 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.404999 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-f9d9-account-create-update-868dt"] Mar 09 18:46:17 crc kubenswrapper[4750]: W0309 18:46:17.410454 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c0825fd_c53a_46cc_9b8a_fca000620d76.slice/crio-e4078c5e31aecef0f8dec1f8fce75573162b90f662043f4372e9d18ed64eadec WatchSource:0}: Error finding container e4078c5e31aecef0f8dec1f8fce75573162b90f662043f4372e9d18ed64eadec: Status 404 returned error can't find the container with id e4078c5e31aecef0f8dec1f8fce75573162b90f662043f4372e9d18ed64eadec Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.679465 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3236a6fb-c288-441c-96e5-7941f818b0af","Type":"ContainerStarted","Data":"ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6"} Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.680690 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.683527 4750 generic.go:334] "Generic (PLEG): container finished" podID="5bf233bd-79f7-4a07-8328-e7491ebb967b" containerID="0a0acb2e286349de69f28fd4e370cbbf4a675cc106e8ee016b6370abdaeac43c" exitCode=0 Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.683588 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p5r85" event={"ID":"5bf233bd-79f7-4a07-8328-e7491ebb967b","Type":"ContainerDied","Data":"0a0acb2e286349de69f28fd4e370cbbf4a675cc106e8ee016b6370abdaeac43c"} Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.687956 4750 generic.go:334] "Generic (PLEG): container finished" podID="567fa16f-93f1-430f-a041-aba3aaec9957" containerID="af8da8c525aa15597d26252e499384ac9131562117fee699016088d024d65f34" exitCode=0 Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.688024 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1af5-account-create-update-c74fv" event={"ID":"567fa16f-93f1-430f-a041-aba3aaec9957","Type":"ContainerDied","Data":"af8da8c525aa15597d26252e499384ac9131562117fee699016088d024d65f34"} Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.692651 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f9d9-account-create-update-868dt" event={"ID":"1c0825fd-c53a-46cc-9b8a-fca000620d76","Type":"ContainerStarted","Data":"fa5afb0335b88ea964b7f114e452dfd93c4085a1cf39e4d92754ed47642d3ccd"} Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.692685 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f9d9-account-create-update-868dt" event={"ID":"1c0825fd-c53a-46cc-9b8a-fca000620d76","Type":"ContainerStarted","Data":"e4078c5e31aecef0f8dec1f8fce75573162b90f662043f4372e9d18ed64eadec"} Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.710754 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=52.882445398 podStartE2EDuration="1m8.71073754s" podCreationTimestamp="2026-03-09 18:45:09 +0000 UTC" firstStartedPulling="2026-03-09 18:45:25.553571316 +0000 UTC m=+1206.896043724" lastFinishedPulling="2026-03-09 18:45:41.381863468 +0000 UTC m=+1222.724335866" observedRunningTime="2026-03-09 18:46:17.706811894 +0000 UTC m=+1259.049284312" watchObservedRunningTime="2026-03-09 18:46:17.71073754 +0000 UTC m=+1259.053209938" Mar 09 18:46:17 crc kubenswrapper[4750]: W0309 18:46:17.738403 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4601d20_845b_4ab0_9f53_0e10aa1b46a7.slice/crio-9e068af8ecb34b6f55b047dcbde8c677f613963fdb0bba956c2c68e5567d5929 WatchSource:0}: Error finding container 9e068af8ecb34b6f55b047dcbde8c677f613963fdb0bba956c2c68e5567d5929: Status 404 returned error can't find the container with id 9e068af8ecb34b6f55b047dcbde8c677f613963fdb0bba956c2c68e5567d5929 Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.739387 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-qlj7j"] Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.740468 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-f9d9-account-create-update-868dt" podStartSLOduration=1.740440505 podStartE2EDuration="1.740440505s" podCreationTimestamp="2026-03-09 18:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:17.729803627 +0000 UTC m=+1259.072276015" watchObservedRunningTime="2026-03-09 18:46:17.740440505 +0000 UTC m=+1259.082912903" Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.824305 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:17 crc kubenswrapper[4750]: I0309 18:46:17.827568 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:18 crc kubenswrapper[4750]: E0309 18:46:18.095695 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c0825fd_c53a_46cc_9b8a_fca000620d76.slice/crio-conmon-fa5afb0335b88ea964b7f114e452dfd93c4085a1cf39e4d92754ed47642d3ccd.scope\": RecentStats: unable to find data in memory cache]" Mar 09 18:46:18 crc kubenswrapper[4750]: I0309 18:46:18.704138 4750 generic.go:334] "Generic (PLEG): container finished" podID="1c0825fd-c53a-46cc-9b8a-fca000620d76" containerID="fa5afb0335b88ea964b7f114e452dfd93c4085a1cf39e4d92754ed47642d3ccd" exitCode=0 Mar 09 18:46:18 crc kubenswrapper[4750]: I0309 18:46:18.704215 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f9d9-account-create-update-868dt" event={"ID":"1c0825fd-c53a-46cc-9b8a-fca000620d76","Type":"ContainerDied","Data":"fa5afb0335b88ea964b7f114e452dfd93c4085a1cf39e4d92754ed47642d3ccd"} Mar 09 18:46:18 crc kubenswrapper[4750]: I0309 18:46:18.710102 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4601d20-845b-4ab0-9f53-0e10aa1b46a7" containerID="f144559cfb61a2b80f797c7a32494d41b8d4cc0ce2b5220a351212b31cef59bf" exitCode=0 Mar 09 18:46:18 crc kubenswrapper[4750]: I0309 18:46:18.710203 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-qlj7j" event={"ID":"f4601d20-845b-4ab0-9f53-0e10aa1b46a7","Type":"ContainerDied","Data":"f144559cfb61a2b80f797c7a32494d41b8d4cc0ce2b5220a351212b31cef59bf"} Mar 09 18:46:18 crc kubenswrapper[4750]: I0309 18:46:18.710285 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-qlj7j" event={"ID":"f4601d20-845b-4ab0-9f53-0e10aa1b46a7","Type":"ContainerStarted","Data":"9e068af8ecb34b6f55b047dcbde8c677f613963fdb0bba956c2c68e5567d5929"} Mar 09 18:46:18 crc kubenswrapper[4750]: I0309 18:46:18.712436 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.131423 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-lr9tk"] Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.133552 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.156130 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lr9tk"] Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.212167 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v9p5\" (UniqueName: \"kubernetes.io/projected/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-kube-api-access-2v9p5\") pod \"glance-db-create-lr9tk\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.212258 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-operator-scripts\") pod \"glance-db-create-lr9tk\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.248828 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-d55d-account-create-update-rx5b2"] Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.250227 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.251759 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.256010 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d55d-account-create-update-rx5b2"] Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.297441 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.313722 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-operator-scripts\") pod \"glance-db-create-lr9tk\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.313820 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9rqn\" (UniqueName: \"kubernetes.io/projected/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-kube-api-access-p9rqn\") pod \"glance-d55d-account-create-update-rx5b2\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.313857 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-operator-scripts\") pod \"glance-d55d-account-create-update-rx5b2\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.313979 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v9p5\" (UniqueName: \"kubernetes.io/projected/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-kube-api-access-2v9p5\") pod \"glance-db-create-lr9tk\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.315375 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-operator-scripts\") pod \"glance-db-create-lr9tk\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.353968 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v9p5\" (UniqueName: \"kubernetes.io/projected/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-kube-api-access-2v9p5\") pod \"glance-db-create-lr9tk\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.415450 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/567fa16f-93f1-430f-a041-aba3aaec9957-operator-scripts\") pod \"567fa16f-93f1-430f-a041-aba3aaec9957\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.415509 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqssj\" (UniqueName: \"kubernetes.io/projected/567fa16f-93f1-430f-a041-aba3aaec9957-kube-api-access-hqssj\") pod \"567fa16f-93f1-430f-a041-aba3aaec9957\" (UID: \"567fa16f-93f1-430f-a041-aba3aaec9957\") " Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.415771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9rqn\" (UniqueName: \"kubernetes.io/projected/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-kube-api-access-p9rqn\") pod \"glance-d55d-account-create-update-rx5b2\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.415799 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-operator-scripts\") pod \"glance-d55d-account-create-update-rx5b2\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.416759 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-operator-scripts\") pod \"glance-d55d-account-create-update-rx5b2\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.417071 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567fa16f-93f1-430f-a041-aba3aaec9957-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "567fa16f-93f1-430f-a041-aba3aaec9957" (UID: "567fa16f-93f1-430f-a041-aba3aaec9957"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.419751 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567fa16f-93f1-430f-a041-aba3aaec9957-kube-api-access-hqssj" (OuterVolumeSpecName: "kube-api-access-hqssj") pod "567fa16f-93f1-430f-a041-aba3aaec9957" (UID: "567fa16f-93f1-430f-a041-aba3aaec9957"). InnerVolumeSpecName "kube-api-access-hqssj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.435029 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9rqn\" (UniqueName: \"kubernetes.io/projected/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-kube-api-access-p9rqn\") pod \"glance-d55d-account-create-update-rx5b2\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.484854 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.486312 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p5r85" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.517689 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/567fa16f-93f1-430f-a041-aba3aaec9957-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.517937 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqssj\" (UniqueName: \"kubernetes.io/projected/567fa16f-93f1-430f-a041-aba3aaec9957-kube-api-access-hqssj\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.576205 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.619016 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf233bd-79f7-4a07-8328-e7491ebb967b-operator-scripts\") pod \"5bf233bd-79f7-4a07-8328-e7491ebb967b\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.619790 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bplsz\" (UniqueName: \"kubernetes.io/projected/5bf233bd-79f7-4a07-8328-e7491ebb967b-kube-api-access-bplsz\") pod \"5bf233bd-79f7-4a07-8328-e7491ebb967b\" (UID: \"5bf233bd-79f7-4a07-8328-e7491ebb967b\") " Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.620297 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bf233bd-79f7-4a07-8328-e7491ebb967b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5bf233bd-79f7-4a07-8328-e7491ebb967b" (UID: "5bf233bd-79f7-4a07-8328-e7491ebb967b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.627907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bf233bd-79f7-4a07-8328-e7491ebb967b-kube-api-access-bplsz" (OuterVolumeSpecName: "kube-api-access-bplsz") pod "5bf233bd-79f7-4a07-8328-e7491ebb967b" (UID: "5bf233bd-79f7-4a07-8328-e7491ebb967b"). InnerVolumeSpecName "kube-api-access-bplsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.723483 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bplsz\" (UniqueName: \"kubernetes.io/projected/5bf233bd-79f7-4a07-8328-e7491ebb967b-kube-api-access-bplsz\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.723506 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bf233bd-79f7-4a07-8328-e7491ebb967b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.727424 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1af5-account-create-update-c74fv" event={"ID":"567fa16f-93f1-430f-a041-aba3aaec9957","Type":"ContainerDied","Data":"be1b7db7c1b4626d0e4e108a3637126126750d4cb941fd2e409605c4418e01fe"} Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.727456 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be1b7db7c1b4626d0e4e108a3637126126750d4cb941fd2e409605c4418e01fe" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.727514 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1af5-account-create-update-c74fv" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.731555 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p5r85" Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.733335 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p5r85" event={"ID":"5bf233bd-79f7-4a07-8328-e7491ebb967b","Type":"ContainerDied","Data":"a78485dfa29d62f8c3ba528b4eb2cce15aaa1eafd9d0aa5b22f070d3f80225bd"} Mar 09 18:46:19 crc kubenswrapper[4750]: I0309 18:46:19.733387 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a78485dfa29d62f8c3ba528b4eb2cce15aaa1eafd9d0aa5b22f070d3f80225bd" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.075777 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d55d-account-create-update-rx5b2"] Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.092895 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.310723 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.316696 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.379138 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lr9tk"] Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.547943 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbp79\" (UniqueName: \"kubernetes.io/projected/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-kube-api-access-mbp79\") pod \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.548159 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-operator-scripts\") pod \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\" (UID: \"f4601d20-845b-4ab0-9f53-0e10aa1b46a7\") " Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.548258 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c0825fd-c53a-46cc-9b8a-fca000620d76-operator-scripts\") pod \"1c0825fd-c53a-46cc-9b8a-fca000620d76\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.548375 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-597nb\" (UniqueName: \"kubernetes.io/projected/1c0825fd-c53a-46cc-9b8a-fca000620d76-kube-api-access-597nb\") pod \"1c0825fd-c53a-46cc-9b8a-fca000620d76\" (UID: \"1c0825fd-c53a-46cc-9b8a-fca000620d76\") " Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.548823 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f4601d20-845b-4ab0-9f53-0e10aa1b46a7" (UID: "f4601d20-845b-4ab0-9f53-0e10aa1b46a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.548857 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c0825fd-c53a-46cc-9b8a-fca000620d76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c0825fd-c53a-46cc-9b8a-fca000620d76" (UID: "1c0825fd-c53a-46cc-9b8a-fca000620d76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.549517 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.549543 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c0825fd-c53a-46cc-9b8a-fca000620d76-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.553970 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-kube-api-access-mbp79" (OuterVolumeSpecName: "kube-api-access-mbp79") pod "f4601d20-845b-4ab0-9f53-0e10aa1b46a7" (UID: "f4601d20-845b-4ab0-9f53-0e10aa1b46a7"). InnerVolumeSpecName "kube-api-access-mbp79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.555006 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c0825fd-c53a-46cc-9b8a-fca000620d76-kube-api-access-597nb" (OuterVolumeSpecName: "kube-api-access-597nb") pod "1c0825fd-c53a-46cc-9b8a-fca000620d76" (UID: "1c0825fd-c53a-46cc-9b8a-fca000620d76"). InnerVolumeSpecName "kube-api-access-597nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.650749 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-597nb\" (UniqueName: \"kubernetes.io/projected/1c0825fd-c53a-46cc-9b8a-fca000620d76-kube-api-access-597nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.651112 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbp79\" (UniqueName: \"kubernetes.io/projected/f4601d20-845b-4ab0-9f53-0e10aa1b46a7-kube-api-access-mbp79\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.750572 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lr9tk" event={"ID":"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20","Type":"ContainerStarted","Data":"5bf3cee0e9565617eecc6eb8fdf568363527ad515dd9eab1ccdab747319a3e42"} Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.750653 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lr9tk" event={"ID":"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20","Type":"ContainerStarted","Data":"1cbf7eeb57f1e48a272960165ed7ddf512a5ebb99abca17eb24f8fef02aafb40"} Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.753218 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d55d-account-create-update-rx5b2" event={"ID":"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7","Type":"ContainerStarted","Data":"38d9b0b593126dc4246880798f9c506666e03f01b7e244012505849cadf74b0a"} Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.753276 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d55d-account-create-update-rx5b2" event={"ID":"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7","Type":"ContainerStarted","Data":"b70cf3342291993fa77c2e7ba877b33ea8e35d1f5d7991571730719ef476ac9b"} Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.756964 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-f9d9-account-create-update-868dt" event={"ID":"1c0825fd-c53a-46cc-9b8a-fca000620d76","Type":"ContainerDied","Data":"e4078c5e31aecef0f8dec1f8fce75573162b90f662043f4372e9d18ed64eadec"} Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.757008 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4078c5e31aecef0f8dec1f8fce75573162b90f662043f4372e9d18ed64eadec" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.757089 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-f9d9-account-create-update-868dt" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.761833 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-qlj7j" event={"ID":"f4601d20-845b-4ab0-9f53-0e10aa1b46a7","Type":"ContainerDied","Data":"9e068af8ecb34b6f55b047dcbde8c677f613963fdb0bba956c2c68e5567d5929"} Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.761890 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-qlj7j" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.761907 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e068af8ecb34b6f55b047dcbde8c677f613963fdb0bba956c2c68e5567d5929" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.778328 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-lr9tk" podStartSLOduration=1.7783039710000001 podStartE2EDuration="1.778303971s" podCreationTimestamp="2026-03-09 18:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:20.774222901 +0000 UTC m=+1262.116695309" watchObservedRunningTime="2026-03-09 18:46:20.778303971 +0000 UTC m=+1262.120776369" Mar 09 18:46:20 crc kubenswrapper[4750]: I0309 18:46:20.795726 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-d55d-account-create-update-rx5b2" podStartSLOduration=1.795703193 podStartE2EDuration="1.795703193s" podCreationTimestamp="2026-03-09 18:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:20.791118969 +0000 UTC m=+1262.133591367" watchObservedRunningTime="2026-03-09 18:46:20.795703193 +0000 UTC m=+1262.138175591" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006328 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-8sxxd"] Mar 09 18:46:21 crc kubenswrapper[4750]: E0309 18:46:21.006673 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4601d20-845b-4ab0-9f53-0e10aa1b46a7" containerName="mariadb-database-create" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006688 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4601d20-845b-4ab0-9f53-0e10aa1b46a7" containerName="mariadb-database-create" Mar 09 18:46:21 crc kubenswrapper[4750]: E0309 18:46:21.006714 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567fa16f-93f1-430f-a041-aba3aaec9957" containerName="mariadb-account-create-update" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006721 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="567fa16f-93f1-430f-a041-aba3aaec9957" containerName="mariadb-account-create-update" Mar 09 18:46:21 crc kubenswrapper[4750]: E0309 18:46:21.006732 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf233bd-79f7-4a07-8328-e7491ebb967b" containerName="mariadb-database-create" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006738 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf233bd-79f7-4a07-8328-e7491ebb967b" containerName="mariadb-database-create" Mar 09 18:46:21 crc kubenswrapper[4750]: E0309 18:46:21.006746 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c0825fd-c53a-46cc-9b8a-fca000620d76" containerName="mariadb-account-create-update" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006752 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c0825fd-c53a-46cc-9b8a-fca000620d76" containerName="mariadb-account-create-update" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006913 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4601d20-845b-4ab0-9f53-0e10aa1b46a7" containerName="mariadb-database-create" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006930 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="567fa16f-93f1-430f-a041-aba3aaec9957" containerName="mariadb-account-create-update" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006940 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c0825fd-c53a-46cc-9b8a-fca000620d76" containerName="mariadb-account-create-update" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.006950 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf233bd-79f7-4a07-8328-e7491ebb967b" containerName="mariadb-database-create" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.007466 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.010511 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.038836 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-8sxxd"] Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.064744 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/166f799b-de40-49b2-bad5-eb95656127d8-operator-scripts\") pod \"root-account-create-update-8sxxd\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.064841 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh6cr\" (UniqueName: \"kubernetes.io/projected/166f799b-de40-49b2-bad5-eb95656127d8-kube-api-access-nh6cr\") pod \"root-account-create-update-8sxxd\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.144775 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.145210 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="prometheus" containerID="cri-o://fdc04f5cb47fe3f3220d52e717ae63c34c4ac331da7289cd7951365dbf2bfdde" gracePeriod=600 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.145703 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="thanos-sidecar" containerID="cri-o://9e6e56cc11a17a86f0ea07eeea6eaa87b9ce12622f50dc0e6e792d2f5213dff5" gracePeriod=600 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.145766 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="config-reloader" containerID="cri-o://aa74b7401b76e1275fb507277ec22d2dff67430bfb85f3824ad625b656f40b80" gracePeriod=600 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.167376 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/166f799b-de40-49b2-bad5-eb95656127d8-operator-scripts\") pod \"root-account-create-update-8sxxd\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.167464 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh6cr\" (UniqueName: \"kubernetes.io/projected/166f799b-de40-49b2-bad5-eb95656127d8-kube-api-access-nh6cr\") pod \"root-account-create-update-8sxxd\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.168202 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/166f799b-de40-49b2-bad5-eb95656127d8-operator-scripts\") pod \"root-account-create-update-8sxxd\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.190235 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh6cr\" (UniqueName: \"kubernetes.io/projected/166f799b-de40-49b2-bad5-eb95656127d8-kube-api-access-nh6cr\") pod \"root-account-create-update-8sxxd\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.325039 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.743370 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.743911 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.774787 4750 generic.go:334] "Generic (PLEG): container finished" podID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerID="9e6e56cc11a17a86f0ea07eeea6eaa87b9ce12622f50dc0e6e792d2f5213dff5" exitCode=0 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.774834 4750 generic.go:334] "Generic (PLEG): container finished" podID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerID="aa74b7401b76e1275fb507277ec22d2dff67430bfb85f3824ad625b656f40b80" exitCode=0 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.774845 4750 generic.go:334] "Generic (PLEG): container finished" podID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerID="fdc04f5cb47fe3f3220d52e717ae63c34c4ac331da7289cd7951365dbf2bfdde" exitCode=0 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.774858 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerDied","Data":"9e6e56cc11a17a86f0ea07eeea6eaa87b9ce12622f50dc0e6e792d2f5213dff5"} Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.774914 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerDied","Data":"aa74b7401b76e1275fb507277ec22d2dff67430bfb85f3824ad625b656f40b80"} Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.774928 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerDied","Data":"fdc04f5cb47fe3f3220d52e717ae63c34c4ac331da7289cd7951365dbf2bfdde"} Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.777366 4750 generic.go:334] "Generic (PLEG): container finished" podID="9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7" containerID="38d9b0b593126dc4246880798f9c506666e03f01b7e244012505849cadf74b0a" exitCode=0 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.777444 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d55d-account-create-update-rx5b2" event={"ID":"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7","Type":"ContainerDied","Data":"38d9b0b593126dc4246880798f9c506666e03f01b7e244012505849cadf74b0a"} Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.779451 4750 generic.go:334] "Generic (PLEG): container finished" podID="eaedb9df-9a83-4ef4-bb62-8fac20a5cb20" containerID="5bf3cee0e9565617eecc6eb8fdf568363527ad515dd9eab1ccdab747319a3e42" exitCode=0 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.779515 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lr9tk" event={"ID":"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20","Type":"ContainerDied","Data":"5bf3cee0e9565617eecc6eb8fdf568363527ad515dd9eab1ccdab747319a3e42"} Mar 09 18:46:21 crc kubenswrapper[4750]: W0309 18:46:21.906345 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod166f799b_de40_49b2_bad5_eb95656127d8.slice/crio-a47d9a3c193e87e974894d65bc318cdfa6a798298b2be23d47af17251ae38006 WatchSource:0}: Error finding container a47d9a3c193e87e974894d65bc318cdfa6a798298b2be23d47af17251ae38006: Status 404 returned error can't find the container with id a47d9a3c193e87e974894d65bc318cdfa6a798298b2be23d47af17251ae38006 Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.906435 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-8sxxd"] Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.945328 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.984713 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92888bd3-e3db-4127-97d2-37801bbbf1df-config-out\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.984920 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-web-config\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.984946 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-2\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.984964 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-1\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.984990 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-config\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985030 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-0\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985064 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9fng\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-kube-api-access-t9fng\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985094 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-thanos-prometheus-http-client-file\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985160 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-tls-assets\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985255 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"92888bd3-e3db-4127-97d2-37801bbbf1df\" (UID: \"92888bd3-e3db-4127-97d2-37801bbbf1df\") " Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985480 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985724 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985758 4750 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.985986 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.989801 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-config" (OuterVolumeSpecName: "config") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.990423 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.992048 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-kube-api-access-t9fng" (OuterVolumeSpecName: "kube-api-access-t9fng") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "kube-api-access-t9fng". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:21 crc kubenswrapper[4750]: I0309 18:46:21.994221 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92888bd3-e3db-4127-97d2-37801bbbf1df-config-out" (OuterVolumeSpecName: "config-out") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.005198 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.022055 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "pvc-a0a46028-c926-4a36-80e0-af0eced5e620". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.025775 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-web-config" (OuterVolumeSpecName: "web-config") pod "92888bd3-e3db-4127-97d2-37801bbbf1df" (UID: "92888bd3-e3db-4127-97d2-37801bbbf1df"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.088519 4750 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-web-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089047 4750 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089119 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089179 4750 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92888bd3-e3db-4127-97d2-37801bbbf1df-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089247 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9fng\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-kube-api-access-t9fng\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089306 4750 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92888bd3-e3db-4127-97d2-37801bbbf1df-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089371 4750 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92888bd3-e3db-4127-97d2-37801bbbf1df-tls-assets\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089548 4750 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") on node \"crc\" " Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.089608 4750 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92888bd3-e3db-4127-97d2-37801bbbf1df-config-out\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.116887 4750 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.117152 4750 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a0a46028-c926-4a36-80e0-af0eced5e620" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620") on node "crc" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.191518 4750 reconciler_common.go:293] "Volume detached for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.797098 4750 generic.go:334] "Generic (PLEG): container finished" podID="166f799b-de40-49b2-bad5-eb95656127d8" containerID="7e8b256fe6a419bbbdb8af0b37473cc0f0e8beb045d44fc31abc6733df215763" exitCode=0 Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.797518 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8sxxd" event={"ID":"166f799b-de40-49b2-bad5-eb95656127d8","Type":"ContainerDied","Data":"7e8b256fe6a419bbbdb8af0b37473cc0f0e8beb045d44fc31abc6733df215763"} Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.797559 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8sxxd" event={"ID":"166f799b-de40-49b2-bad5-eb95656127d8","Type":"ContainerStarted","Data":"a47d9a3c193e87e974894d65bc318cdfa6a798298b2be23d47af17251ae38006"} Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.800295 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.800304 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92888bd3-e3db-4127-97d2-37801bbbf1df","Type":"ContainerDied","Data":"faad1e19f72a60152ea7f79c423bfeb6880334dde764c0e096c2618519521912"} Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.800593 4750 scope.go:117] "RemoveContainer" containerID="9e6e56cc11a17a86f0ea07eeea6eaa87b9ce12622f50dc0e6e792d2f5213dff5" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.852324 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.872852 4750 scope.go:117] "RemoveContainer" containerID="aa74b7401b76e1275fb507277ec22d2dff67430bfb85f3824ad625b656f40b80" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.872886 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884175 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:46:22 crc kubenswrapper[4750]: E0309 18:46:22.884567 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="init-config-reloader" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884585 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="init-config-reloader" Mar 09 18:46:22 crc kubenswrapper[4750]: E0309 18:46:22.884599 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="thanos-sidecar" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884606 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="thanos-sidecar" Mar 09 18:46:22 crc kubenswrapper[4750]: E0309 18:46:22.884624 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="prometheus" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884643 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="prometheus" Mar 09 18:46:22 crc kubenswrapper[4750]: E0309 18:46:22.884663 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="config-reloader" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884669 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="config-reloader" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884816 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="prometheus" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884832 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="config-reloader" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.884843 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" containerName="thanos-sidecar" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.886328 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.888313 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.896951 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.898110 4750 scope.go:117] "RemoveContainer" containerID="fdc04f5cb47fe3f3220d52e717ae63c34c4ac331da7289cd7951365dbf2bfdde" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.917967 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-d8hlx" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.918225 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.919789 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.919897 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.920114 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.920185 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.930046 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.934915 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 09 18:46:22 crc kubenswrapper[4750]: I0309 18:46:22.968685 4750 scope.go:117] "RemoveContainer" containerID="1624d9c60026bcd3e01d587842fe8b34482e04207adfcf064ea1215fd2120205" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.019720 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrlzl\" (UniqueName: \"kubernetes.io/projected/6ca49786-3511-4719-90de-963bafdd8429-kube-api-access-zrlzl\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.019784 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.019817 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.019877 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.019911 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.019938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.019985 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ca49786-3511-4719-90de-963bafdd8429-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.020239 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.020333 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ca49786-3511-4719-90de-963bafdd8429-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.020369 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.020441 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.020543 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.020588 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.121933 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122004 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122029 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122060 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ca49786-3511-4719-90de-963bafdd8429-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122083 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122108 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ca49786-3511-4719-90de-963bafdd8429-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122131 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122167 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122219 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122248 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122269 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrlzl\" (UniqueName: \"kubernetes.io/projected/6ca49786-3511-4719-90de-963bafdd8429-kube-api-access-zrlzl\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122289 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.122309 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.123113 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.123521 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.126369 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/6ca49786-3511-4719-90de-963bafdd8429-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.126989 4750 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.127059 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c80612d5a9520b4a0639dd996533a4e971f0dc6e01d511a4a320a863e8837823/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.130686 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.131222 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ca49786-3511-4719-90de-963bafdd8429-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.131552 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.131687 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.134316 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.134862 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.138369 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6ca49786-3511-4719-90de-963bafdd8429-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.142764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ca49786-3511-4719-90de-963bafdd8429-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.155722 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrlzl\" (UniqueName: \"kubernetes.io/projected/6ca49786-3511-4719-90de-963bafdd8429-kube-api-access-zrlzl\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.231083 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a0a46028-c926-4a36-80e0-af0eced5e620\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0a46028-c926-4a36-80e0-af0eced5e620\") pod \"prometheus-metric-storage-0\" (UID: \"6ca49786-3511-4719-90de-963bafdd8429\") " pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.243520 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.261203 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.325136 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-operator-scripts\") pod \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.325233 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v9p5\" (UniqueName: \"kubernetes.io/projected/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-kube-api-access-2v9p5\") pod \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\" (UID: \"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20\") " Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.327656 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eaedb9df-9a83-4ef4-bb62-8fac20a5cb20" (UID: "eaedb9df-9a83-4ef4-bb62-8fac20a5cb20"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.329728 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.330901 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-kube-api-access-2v9p5" (OuterVolumeSpecName: "kube-api-access-2v9p5") pod "eaedb9df-9a83-4ef4-bb62-8fac20a5cb20" (UID: "eaedb9df-9a83-4ef4-bb62-8fac20a5cb20"). InnerVolumeSpecName "kube-api-access-2v9p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.374663 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.385449 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92888bd3-e3db-4127-97d2-37801bbbf1df" path="/var/lib/kubelet/pods/92888bd3-e3db-4127-97d2-37801bbbf1df/volumes" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.430988 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v9p5\" (UniqueName: \"kubernetes.io/projected/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20-kube-api-access-2v9p5\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.532703 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-operator-scripts\") pod \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.533057 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9rqn\" (UniqueName: \"kubernetes.io/projected/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-kube-api-access-p9rqn\") pod \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\" (UID: \"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7\") " Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.533539 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7" (UID: "9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.533866 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.539063 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-kube-api-access-p9rqn" (OuterVolumeSpecName: "kube-api-access-p9rqn") pod "9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7" (UID: "9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7"). InnerVolumeSpecName "kube-api-access-p9rqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.635447 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9rqn\" (UniqueName: \"kubernetes.io/projected/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7-kube-api-access-p9rqn\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.812277 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d55d-account-create-update-rx5b2" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.812268 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d55d-account-create-update-rx5b2" event={"ID":"9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7","Type":"ContainerDied","Data":"b70cf3342291993fa77c2e7ba877b33ea8e35d1f5d7991571730719ef476ac9b"} Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.812441 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b70cf3342291993fa77c2e7ba877b33ea8e35d1f5d7991571730719ef476ac9b" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.814107 4750 generic.go:334] "Generic (PLEG): container finished" podID="3926781f-b03f-45a7-82e5-10abf0c16c1e" containerID="ef16edbd1f978bdde45acc1dde08e21f76e7d8c166b9ac45794ad9861eaa07c0" exitCode=0 Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.814218 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"3926781f-b03f-45a7-82e5-10abf0c16c1e","Type":"ContainerDied","Data":"ef16edbd1f978bdde45acc1dde08e21f76e7d8c166b9ac45794ad9861eaa07c0"} Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.816477 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lr9tk" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.816572 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lr9tk" event={"ID":"eaedb9df-9a83-4ef4-bb62-8fac20a5cb20","Type":"ContainerDied","Data":"1cbf7eeb57f1e48a272960165ed7ddf512a5ebb99abca17eb24f8fef02aafb40"} Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.816639 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cbf7eeb57f1e48a272960165ed7ddf512a5ebb99abca17eb24f8fef02aafb40" Mar 09 18:46:23 crc kubenswrapper[4750]: I0309 18:46:23.848515 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.326023 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-n5qz8" podUID="5e9fd826-3829-4379-9717-1b76ef376a39" containerName="ovn-controller" probeResult="failure" output=< Mar 09 18:46:24 crc kubenswrapper[4750]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 09 18:46:24 crc kubenswrapper[4750]: > Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.355375 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.368229 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6dgcc" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.501303 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.570995 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8k6sd"] Mar 09 18:46:24 crc kubenswrapper[4750]: E0309 18:46:24.571908 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7" containerName="mariadb-account-create-update" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.571933 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7" containerName="mariadb-account-create-update" Mar 09 18:46:24 crc kubenswrapper[4750]: E0309 18:46:24.571953 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166f799b-de40-49b2-bad5-eb95656127d8" containerName="mariadb-account-create-update" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.571966 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="166f799b-de40-49b2-bad5-eb95656127d8" containerName="mariadb-account-create-update" Mar 09 18:46:24 crc kubenswrapper[4750]: E0309 18:46:24.572003 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaedb9df-9a83-4ef4-bb62-8fac20a5cb20" containerName="mariadb-database-create" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.572011 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaedb9df-9a83-4ef4-bb62-8fac20a5cb20" containerName="mariadb-database-create" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.572370 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="166f799b-de40-49b2-bad5-eb95656127d8" containerName="mariadb-account-create-update" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.572386 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaedb9df-9a83-4ef4-bb62-8fac20a5cb20" containerName="mariadb-database-create" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.572420 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7" containerName="mariadb-account-create-update" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.575591 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.583445 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.583839 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qbg7p" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.607956 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8k6sd"] Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.647081 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-n5qz8-config-kkxlh"] Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.648474 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.652119 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.658050 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n5qz8-config-kkxlh"] Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.674931 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh6cr\" (UniqueName: \"kubernetes.io/projected/166f799b-de40-49b2-bad5-eb95656127d8-kube-api-access-nh6cr\") pod \"166f799b-de40-49b2-bad5-eb95656127d8\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.675104 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/166f799b-de40-49b2-bad5-eb95656127d8-operator-scripts\") pod \"166f799b-de40-49b2-bad5-eb95656127d8\" (UID: \"166f799b-de40-49b2-bad5-eb95656127d8\") " Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.675338 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-combined-ca-bundle\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.675393 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-db-sync-config-data\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.675411 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-config-data\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.675444 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfblk\" (UniqueName: \"kubernetes.io/projected/46354cf1-acc8-45a7-a1e7-92f65230b565-kube-api-access-cfblk\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.676169 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/166f799b-de40-49b2-bad5-eb95656127d8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "166f799b-de40-49b2-bad5-eb95656127d8" (UID: "166f799b-de40-49b2-bad5-eb95656127d8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.685824 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/166f799b-de40-49b2-bad5-eb95656127d8-kube-api-access-nh6cr" (OuterVolumeSpecName: "kube-api-access-nh6cr") pod "166f799b-de40-49b2-bad5-eb95656127d8" (UID: "166f799b-de40-49b2-bad5-eb95656127d8"). InnerVolumeSpecName "kube-api-access-nh6cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777137 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777207 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g72s\" (UniqueName: \"kubernetes.io/projected/1c58f537-36f5-4d3f-a50b-305c146100c3-kube-api-access-2g72s\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777312 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-db-sync-config-data\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777366 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-config-data\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777471 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-scripts\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777494 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfblk\" (UniqueName: \"kubernetes.io/projected/46354cf1-acc8-45a7-a1e7-92f65230b565-kube-api-access-cfblk\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-log-ovn\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777681 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run-ovn\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777714 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-additional-scripts\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.777863 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-combined-ca-bundle\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.778430 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/166f799b-de40-49b2-bad5-eb95656127d8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.778469 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh6cr\" (UniqueName: \"kubernetes.io/projected/166f799b-de40-49b2-bad5-eb95656127d8-kube-api-access-nh6cr\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.781417 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-db-sync-config-data\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.782066 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-config-data\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.784726 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-combined-ca-bundle\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.797787 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfblk\" (UniqueName: \"kubernetes.io/projected/46354cf1-acc8-45a7-a1e7-92f65230b565-kube-api-access-cfblk\") pod \"glance-db-sync-8k6sd\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.826485 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"3926781f-b03f-45a7-82e5-10abf0c16c1e","Type":"ContainerStarted","Data":"b689f52c22169de19bcb9ca021db3f72ec790e285b2a7a8429cbef68827655ac"} Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.827745 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.829387 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8sxxd" event={"ID":"166f799b-de40-49b2-bad5-eb95656127d8","Type":"ContainerDied","Data":"a47d9a3c193e87e974894d65bc318cdfa6a798298b2be23d47af17251ae38006"} Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.829415 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a47d9a3c193e87e974894d65bc318cdfa6a798298b2be23d47af17251ae38006" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.829467 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sxxd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.830709 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ca49786-3511-4719-90de-963bafdd8429","Type":"ContainerStarted","Data":"ac5b834cc4b9d513071ae207b78a7ea6fc02f8104999494257cd9486d82eb418"} Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.854131 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/notifications-rabbitmq-server-0" podStartSLOduration=-9223371961.000664 podStartE2EDuration="1m15.854112462s" podCreationTimestamp="2026-03-09 18:45:09 +0000 UTC" firstStartedPulling="2026-03-09 18:45:25.539168286 +0000 UTC m=+1206.881640684" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:24.85143496 +0000 UTC m=+1266.193907368" watchObservedRunningTime="2026-03-09 18:46:24.854112462 +0000 UTC m=+1266.196584860" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.883853 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.883901 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g72s\" (UniqueName: \"kubernetes.io/projected/1c58f537-36f5-4d3f-a50b-305c146100c3-kube-api-access-2g72s\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.883943 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-scripts\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.883977 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-log-ovn\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.884015 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run-ovn\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.884034 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-additional-scripts\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.884349 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-log-ovn\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.884353 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.884409 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run-ovn\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.884857 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-additional-scripts\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.886008 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-scripts\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.907141 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g72s\" (UniqueName: \"kubernetes.io/projected/1c58f537-36f5-4d3f-a50b-305c146100c3-kube-api-access-2g72s\") pod \"ovn-controller-n5qz8-config-kkxlh\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.907822 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:24 crc kubenswrapper[4750]: I0309 18:46:24.965051 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.219608 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-hnrfj"] Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.221989 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.243705 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-10ba-account-create-update-5p6zt"] Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.245347 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.247917 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.258975 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-10ba-account-create-update-5p6zt"] Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.271564 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hnrfj"] Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.397274 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f192f0ed-c217-472f-9535-91518a9337d4-operator-scripts\") pod \"keystone-db-create-hnrfj\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.397333 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th98z\" (UniqueName: \"kubernetes.io/projected/f192f0ed-c217-472f-9535-91518a9337d4-kube-api-access-th98z\") pod \"keystone-db-create-hnrfj\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.397389 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l9bz\" (UniqueName: \"kubernetes.io/projected/33e86472-4600-45ec-888c-988c15731c69-kube-api-access-7l9bz\") pod \"keystone-10ba-account-create-update-5p6zt\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.397426 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e86472-4600-45ec-888c-988c15731c69-operator-scripts\") pod \"keystone-10ba-account-create-update-5p6zt\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.441758 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n5qz8-config-kkxlh"] Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.499174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th98z\" (UniqueName: \"kubernetes.io/projected/f192f0ed-c217-472f-9535-91518a9337d4-kube-api-access-th98z\") pod \"keystone-db-create-hnrfj\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.499252 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l9bz\" (UniqueName: \"kubernetes.io/projected/33e86472-4600-45ec-888c-988c15731c69-kube-api-access-7l9bz\") pod \"keystone-10ba-account-create-update-5p6zt\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.499301 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e86472-4600-45ec-888c-988c15731c69-operator-scripts\") pod \"keystone-10ba-account-create-update-5p6zt\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.499484 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f192f0ed-c217-472f-9535-91518a9337d4-operator-scripts\") pod \"keystone-db-create-hnrfj\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.500238 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e86472-4600-45ec-888c-988c15731c69-operator-scripts\") pod \"keystone-10ba-account-create-update-5p6zt\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.500760 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f192f0ed-c217-472f-9535-91518a9337d4-operator-scripts\") pod \"keystone-db-create-hnrfj\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.523063 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th98z\" (UniqueName: \"kubernetes.io/projected/f192f0ed-c217-472f-9535-91518a9337d4-kube-api-access-th98z\") pod \"keystone-db-create-hnrfj\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.524695 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l9bz\" (UniqueName: \"kubernetes.io/projected/33e86472-4600-45ec-888c-988c15731c69-kube-api-access-7l9bz\") pod \"keystone-10ba-account-create-update-5p6zt\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.542304 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8k6sd"] Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.562389 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.576052 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.848203 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n5qz8-config-kkxlh" event={"ID":"1c58f537-36f5-4d3f-a50b-305c146100c3","Type":"ContainerStarted","Data":"fd9dca0532967b0ddaa90d4cf32ddcfda522cb858db124048657d6f638d9ec6e"} Mar 09 18:46:25 crc kubenswrapper[4750]: I0309 18:46:25.850289 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8k6sd" event={"ID":"46354cf1-acc8-45a7-a1e7-92f65230b565","Type":"ContainerStarted","Data":"b5c483538d89322d5430e91bbbb6463c20bcd8882640341ef75402a997dea45e"} Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.085106 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hnrfj"] Mar 09 18:46:26 crc kubenswrapper[4750]: W0309 18:46:26.096500 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf192f0ed_c217_472f_9535_91518a9337d4.slice/crio-a974660de9111c56837ace038b1a583e14b4b22407862a0d79a492102514507a WatchSource:0}: Error finding container a974660de9111c56837ace038b1a583e14b4b22407862a0d79a492102514507a: Status 404 returned error can't find the container with id a974660de9111c56837ace038b1a583e14b4b22407862a0d79a492102514507a Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.217889 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-10ba-account-create-update-5p6zt"] Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.320379 4750 scope.go:117] "RemoveContainer" containerID="b54aed15db155e225dfc62a44b1737b87c9e40783d39c0fdc434ca62f0803160" Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.861121 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-10ba-account-create-update-5p6zt" event={"ID":"33e86472-4600-45ec-888c-988c15731c69","Type":"ContainerStarted","Data":"2b9d4ca0db520dea4fca24e91369c5920f0ae3eb6c68ee18ca661f9d482ece8e"} Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.870585 4750 generic.go:334] "Generic (PLEG): container finished" podID="1c58f537-36f5-4d3f-a50b-305c146100c3" containerID="1ed9d78873c42f484fd82ac83617e0adac66f1e76adc7b75b939bd6244566400" exitCode=0 Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.870684 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n5qz8-config-kkxlh" event={"ID":"1c58f537-36f5-4d3f-a50b-305c146100c3","Type":"ContainerDied","Data":"1ed9d78873c42f484fd82ac83617e0adac66f1e76adc7b75b939bd6244566400"} Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.877089 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hnrfj" event={"ID":"f192f0ed-c217-472f-9535-91518a9337d4","Type":"ContainerStarted","Data":"3477901d8ff8b9596d5a273aeaf7f08f34e9f6ef08b3a89e9699fd44163c6ce2"} Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.877118 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hnrfj" event={"ID":"f192f0ed-c217-472f-9535-91518a9337d4","Type":"ContainerStarted","Data":"a974660de9111c56837ace038b1a583e14b4b22407862a0d79a492102514507a"} Mar 09 18:46:26 crc kubenswrapper[4750]: I0309 18:46:26.959061 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-hnrfj" podStartSLOduration=1.959025389 podStartE2EDuration="1.959025389s" podCreationTimestamp="2026-03-09 18:46:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:26.940551549 +0000 UTC m=+1268.283023947" watchObservedRunningTime="2026-03-09 18:46:26.959025389 +0000 UTC m=+1268.301497787" Mar 09 18:46:27 crc kubenswrapper[4750]: I0309 18:46:27.529673 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-8sxxd"] Mar 09 18:46:27 crc kubenswrapper[4750]: I0309 18:46:27.536513 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-8sxxd"] Mar 09 18:46:27 crc kubenswrapper[4750]: I0309 18:46:27.895947 4750 generic.go:334] "Generic (PLEG): container finished" podID="f192f0ed-c217-472f-9535-91518a9337d4" containerID="3477901d8ff8b9596d5a273aeaf7f08f34e9f6ef08b3a89e9699fd44163c6ce2" exitCode=0 Mar 09 18:46:27 crc kubenswrapper[4750]: I0309 18:46:27.896265 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hnrfj" event={"ID":"f192f0ed-c217-472f-9535-91518a9337d4","Type":"ContainerDied","Data":"3477901d8ff8b9596d5a273aeaf7f08f34e9f6ef08b3a89e9699fd44163c6ce2"} Mar 09 18:46:27 crc kubenswrapper[4750]: I0309 18:46:27.899364 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ca49786-3511-4719-90de-963bafdd8429","Type":"ContainerStarted","Data":"8b194f22ccbae08a10a136ee11f3cf1be12159dcd76ceef67718aaff8a597ad5"} Mar 09 18:46:27 crc kubenswrapper[4750]: I0309 18:46:27.903312 4750 generic.go:334] "Generic (PLEG): container finished" podID="33e86472-4600-45ec-888c-988c15731c69" containerID="e52a5353a572e957bfb271bd766b9b481f27132b075183bed227dcef209ef206" exitCode=0 Mar 09 18:46:27 crc kubenswrapper[4750]: I0309 18:46:27.903363 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-10ba-account-create-update-5p6zt" event={"ID":"33e86472-4600-45ec-888c-988c15731c69","Type":"ContainerDied","Data":"e52a5353a572e957bfb271bd766b9b481f27132b075183bed227dcef209ef206"} Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.248275 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394095 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run-ovn\") pod \"1c58f537-36f5-4d3f-a50b-305c146100c3\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394233 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g72s\" (UniqueName: \"kubernetes.io/projected/1c58f537-36f5-4d3f-a50b-305c146100c3-kube-api-access-2g72s\") pod \"1c58f537-36f5-4d3f-a50b-305c146100c3\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394261 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-log-ovn\") pod \"1c58f537-36f5-4d3f-a50b-305c146100c3\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394395 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-scripts\") pod \"1c58f537-36f5-4d3f-a50b-305c146100c3\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394424 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1c58f537-36f5-4d3f-a50b-305c146100c3" (UID: "1c58f537-36f5-4d3f-a50b-305c146100c3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394603 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1c58f537-36f5-4d3f-a50b-305c146100c3" (UID: "1c58f537-36f5-4d3f-a50b-305c146100c3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394825 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-additional-scripts\") pod \"1c58f537-36f5-4d3f-a50b-305c146100c3\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.394869 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run\") pod \"1c58f537-36f5-4d3f-a50b-305c146100c3\" (UID: \"1c58f537-36f5-4d3f-a50b-305c146100c3\") " Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.395061 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run" (OuterVolumeSpecName: "var-run") pod "1c58f537-36f5-4d3f-a50b-305c146100c3" (UID: "1c58f537-36f5-4d3f-a50b-305c146100c3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.395509 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1c58f537-36f5-4d3f-a50b-305c146100c3" (UID: "1c58f537-36f5-4d3f-a50b-305c146100c3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.395884 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-scripts" (OuterVolumeSpecName: "scripts") pod "1c58f537-36f5-4d3f-a50b-305c146100c3" (UID: "1c58f537-36f5-4d3f-a50b-305c146100c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.396356 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.396401 4750 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1c58f537-36f5-4d3f-a50b-305c146100c3-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.396416 4750 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.396430 4750 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.396444 4750 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1c58f537-36f5-4d3f-a50b-305c146100c3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.402021 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c58f537-36f5-4d3f-a50b-305c146100c3-kube-api-access-2g72s" (OuterVolumeSpecName: "kube-api-access-2g72s") pod "1c58f537-36f5-4d3f-a50b-305c146100c3" (UID: "1c58f537-36f5-4d3f-a50b-305c146100c3"). InnerVolumeSpecName "kube-api-access-2g72s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.498115 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g72s\" (UniqueName: \"kubernetes.io/projected/1c58f537-36f5-4d3f-a50b-305c146100c3-kube-api-access-2g72s\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.917082 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n5qz8-config-kkxlh" Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.917483 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n5qz8-config-kkxlh" event={"ID":"1c58f537-36f5-4d3f-a50b-305c146100c3","Type":"ContainerDied","Data":"fd9dca0532967b0ddaa90d4cf32ddcfda522cb858db124048657d6f638d9ec6e"} Mar 09 18:46:28 crc kubenswrapper[4750]: I0309 18:46:28.917545 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd9dca0532967b0ddaa90d4cf32ddcfda522cb858db124048657d6f638d9ec6e" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.351233 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.410374 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.428479 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="166f799b-de40-49b2-bad5-eb95656127d8" path="/var/lib/kubelet/pods/166f799b-de40-49b2-bad5-eb95656127d8/volumes" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.454971 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-n5qz8" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.474409 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-n5qz8-config-kkxlh"] Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.511430 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-n5qz8-config-kkxlh"] Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.528060 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e86472-4600-45ec-888c-988c15731c69-operator-scripts\") pod \"33e86472-4600-45ec-888c-988c15731c69\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.528204 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th98z\" (UniqueName: \"kubernetes.io/projected/f192f0ed-c217-472f-9535-91518a9337d4-kube-api-access-th98z\") pod \"f192f0ed-c217-472f-9535-91518a9337d4\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.528284 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l9bz\" (UniqueName: \"kubernetes.io/projected/33e86472-4600-45ec-888c-988c15731c69-kube-api-access-7l9bz\") pod \"33e86472-4600-45ec-888c-988c15731c69\" (UID: \"33e86472-4600-45ec-888c-988c15731c69\") " Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.528387 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f192f0ed-c217-472f-9535-91518a9337d4-operator-scripts\") pod \"f192f0ed-c217-472f-9535-91518a9337d4\" (UID: \"f192f0ed-c217-472f-9535-91518a9337d4\") " Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.528503 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33e86472-4600-45ec-888c-988c15731c69-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "33e86472-4600-45ec-888c-988c15731c69" (UID: "33e86472-4600-45ec-888c-988c15731c69"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.528957 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f192f0ed-c217-472f-9535-91518a9337d4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f192f0ed-c217-472f-9535-91518a9337d4" (UID: "f192f0ed-c217-472f-9535-91518a9337d4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.529007 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e86472-4600-45ec-888c-988c15731c69-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.536999 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e86472-4600-45ec-888c-988c15731c69-kube-api-access-7l9bz" (OuterVolumeSpecName: "kube-api-access-7l9bz") pod "33e86472-4600-45ec-888c-988c15731c69" (UID: "33e86472-4600-45ec-888c-988c15731c69"). InnerVolumeSpecName "kube-api-access-7l9bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.557743 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f192f0ed-c217-472f-9535-91518a9337d4-kube-api-access-th98z" (OuterVolumeSpecName: "kube-api-access-th98z") pod "f192f0ed-c217-472f-9535-91518a9337d4" (UID: "f192f0ed-c217-472f-9535-91518a9337d4"). InnerVolumeSpecName "kube-api-access-th98z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.632698 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l9bz\" (UniqueName: \"kubernetes.io/projected/33e86472-4600-45ec-888c-988c15731c69-kube-api-access-7l9bz\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.632731 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f192f0ed-c217-472f-9535-91518a9337d4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.632743 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th98z\" (UniqueName: \"kubernetes.io/projected/f192f0ed-c217-472f-9535-91518a9337d4-kube-api-access-th98z\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.836151 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.841167 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd12d247-7fd6-468d-8450-c395c4ee57ac-etc-swift\") pod \"swift-storage-0\" (UID: \"cd12d247-7fd6-468d-8450-c395c4ee57ac\") " pod="openstack/swift-storage-0" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.926328 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-10ba-account-create-update-5p6zt" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.926324 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-10ba-account-create-update-5p6zt" event={"ID":"33e86472-4600-45ec-888c-988c15731c69","Type":"ContainerDied","Data":"2b9d4ca0db520dea4fca24e91369c5920f0ae3eb6c68ee18ca661f9d482ece8e"} Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.926376 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b9d4ca0db520dea4fca24e91369c5920f0ae3eb6c68ee18ca661f9d482ece8e" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.928685 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hnrfj" event={"ID":"f192f0ed-c217-472f-9535-91518a9337d4","Type":"ContainerDied","Data":"a974660de9111c56837ace038b1a583e14b4b22407862a0d79a492102514507a"} Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.928709 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hnrfj" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.928711 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a974660de9111c56837ace038b1a583e14b4b22407862a0d79a492102514507a" Mar 09 18:46:29 crc kubenswrapper[4750]: I0309 18:46:29.987922 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:30.526707 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="378d9b58-f830-4d61-b408-26668a301507" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:30.581572 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 09 18:46:31 crc kubenswrapper[4750]: W0309 18:46:30.591137 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd12d247_7fd6_468d_8450_c395c4ee57ac.slice/crio-b725994770d9854348c1249d250b5a4adb8544cd1ddb98de9b184aa437741051 WatchSource:0}: Error finding container b725994770d9854348c1249d250b5a4adb8544cd1ddb98de9b184aa437741051: Status 404 returned error can't find the container with id b725994770d9854348c1249d250b5a4adb8544cd1ddb98de9b184aa437741051 Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:30.794050 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:30.939741 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"b725994770d9854348c1249d250b5a4adb8544cd1ddb98de9b184aa437741051"} Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:31.387015 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c58f537-36f5-4d3f-a50b-305c146100c3" path="/var/lib/kubelet/pods/1c58f537-36f5-4d3f-a50b-305c146100c3/volumes" Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:31.573937 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:31.954904 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"f843abf6ad22ef98590769371639ab06ab0cb77c4cbc3054f92425e8c8a4ec8b"} Mar 09 18:46:31 crc kubenswrapper[4750]: I0309 18:46:31.954983 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"05e127e5c38b9798044aefdc3b9fd9d81da0436a3f92ce785283cefbdc908025"} Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.555087 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-mqfgs"] Mar 09 18:46:32 crc kubenswrapper[4750]: E0309 18:46:32.556116 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c58f537-36f5-4d3f-a50b-305c146100c3" containerName="ovn-config" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.556139 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c58f537-36f5-4d3f-a50b-305c146100c3" containerName="ovn-config" Mar 09 18:46:32 crc kubenswrapper[4750]: E0309 18:46:32.556177 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f192f0ed-c217-472f-9535-91518a9337d4" containerName="mariadb-database-create" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.556185 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f192f0ed-c217-472f-9535-91518a9337d4" containerName="mariadb-database-create" Mar 09 18:46:32 crc kubenswrapper[4750]: E0309 18:46:32.556198 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e86472-4600-45ec-888c-988c15731c69" containerName="mariadb-account-create-update" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.556204 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e86472-4600-45ec-888c-988c15731c69" containerName="mariadb-account-create-update" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.556414 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e86472-4600-45ec-888c-988c15731c69" containerName="mariadb-account-create-update" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.556431 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f192f0ed-c217-472f-9535-91518a9337d4" containerName="mariadb-database-create" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.556442 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c58f537-36f5-4d3f-a50b-305c146100c3" containerName="ovn-config" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.561534 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.565676 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.569327 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-mqfgs"] Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.706173 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6nn2\" (UniqueName: \"kubernetes.io/projected/783ff672-9bd9-4092-8823-9a7bc9d2364a-kube-api-access-t6nn2\") pod \"root-account-create-update-mqfgs\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.706721 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/783ff672-9bd9-4092-8823-9a7bc9d2364a-operator-scripts\") pod \"root-account-create-update-mqfgs\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.808796 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/783ff672-9bd9-4092-8823-9a7bc9d2364a-operator-scripts\") pod \"root-account-create-update-mqfgs\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.808940 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6nn2\" (UniqueName: \"kubernetes.io/projected/783ff672-9bd9-4092-8823-9a7bc9d2364a-kube-api-access-t6nn2\") pod \"root-account-create-update-mqfgs\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.809709 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/783ff672-9bd9-4092-8823-9a7bc9d2364a-operator-scripts\") pod \"root-account-create-update-mqfgs\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.829495 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6nn2\" (UniqueName: \"kubernetes.io/projected/783ff672-9bd9-4092-8823-9a7bc9d2364a-kube-api-access-t6nn2\") pod \"root-account-create-update-mqfgs\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.934406 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.971895 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"98176983ccd9520e95324f2d385555c849848440a1b5b678a6e96b9211b04589"} Mar 09 18:46:32 crc kubenswrapper[4750]: I0309 18:46:32.971963 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"20501098c92c273fc5675747c79d6ccfabca2d887c652f6308f986d3c4b4d38d"} Mar 09 18:46:33 crc kubenswrapper[4750]: I0309 18:46:33.410811 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-mqfgs"] Mar 09 18:46:33 crc kubenswrapper[4750]: I0309 18:46:33.983489 4750 generic.go:334] "Generic (PLEG): container finished" podID="6ca49786-3511-4719-90de-963bafdd8429" containerID="8b194f22ccbae08a10a136ee11f3cf1be12159dcd76ceef67718aaff8a597ad5" exitCode=0 Mar 09 18:46:33 crc kubenswrapper[4750]: I0309 18:46:33.983550 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ca49786-3511-4719-90de-963bafdd8429","Type":"ContainerDied","Data":"8b194f22ccbae08a10a136ee11f3cf1be12159dcd76ceef67718aaff8a597ad5"} Mar 09 18:46:40 crc kubenswrapper[4750]: W0309 18:46:40.438463 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod783ff672_9bd9_4092_8823_9a7bc9d2364a.slice/crio-fa85a72a45c57ebd0f46a9572e66eed4f7be97d77f0b461864664c48472bcbe6 WatchSource:0}: Error finding container fa85a72a45c57ebd0f46a9572e66eed4f7be97d77f0b461864664c48472bcbe6: Status 404 returned error can't find the container with id fa85a72a45c57ebd0f46a9572e66eed4f7be97d77f0b461864664c48472bcbe6 Mar 09 18:46:40 crc kubenswrapper[4750]: I0309 18:46:40.526404 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 09 18:46:40 crc kubenswrapper[4750]: I0309 18:46:40.792907 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.103622 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"139089ffc4b9fd5ef6f4162dba22ded9d2775e3dbe22d9baf10474a6866332af"} Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.113703 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ca49786-3511-4719-90de-963bafdd8429","Type":"ContainerStarted","Data":"61a036353e9c03d9ae447eccb156d024b563a22214f0671714eb567fccb216ac"} Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.124358 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqfgs" event={"ID":"783ff672-9bd9-4092-8823-9a7bc9d2364a","Type":"ContainerStarted","Data":"f73e1c4aa1476494bae0883ca8680b278dc6dc9f17f0cb23e2ecc770967f6191"} Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.124409 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqfgs" event={"ID":"783ff672-9bd9-4092-8823-9a7bc9d2364a","Type":"ContainerStarted","Data":"fa85a72a45c57ebd0f46a9572e66eed4f7be97d77f0b461864664c48472bcbe6"} Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.147831 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/notifications-rabbitmq-server-0" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.162284 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-mqfgs" podStartSLOduration=9.162257319 podStartE2EDuration="9.162257319s" podCreationTimestamp="2026-03-09 18:46:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:41.16043553 +0000 UTC m=+1282.502907938" watchObservedRunningTime="2026-03-09 18:46:41.162257319 +0000 UTC m=+1282.504729707" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.311497 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rfqnd"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.312969 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.333134 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rfqnd"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.348658 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-ba6b-account-create-update-z9h5p"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.349899 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.356082 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.367149 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gssq6\" (UniqueName: \"kubernetes.io/projected/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-kube-api-access-gssq6\") pod \"barbican-db-create-rfqnd\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.367227 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-operator-scripts\") pod \"barbican-db-create-rfqnd\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.393389 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ba6b-account-create-update-z9h5p"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.450727 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-mzvtm"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.452167 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.465318 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ln7q6"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.466610 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.468912 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k49p\" (UniqueName: \"kubernetes.io/projected/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-kube-api-access-5k49p\") pod \"barbican-ba6b-account-create-update-z9h5p\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.469481 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gssq6\" (UniqueName: \"kubernetes.io/projected/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-kube-api-access-gssq6\") pod \"barbican-db-create-rfqnd\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.469660 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-operator-scripts\") pod \"barbican-db-create-rfqnd\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.469804 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-operator-scripts\") pod \"barbican-ba6b-account-create-update-z9h5p\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.469307 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.469488 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.470622 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-operator-scripts\") pod \"barbican-db-create-rfqnd\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.481455 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.484206 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9flmk" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.496703 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mzvtm"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.517256 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gssq6\" (UniqueName: \"kubernetes.io/projected/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-kube-api-access-gssq6\") pod \"barbican-db-create-rfqnd\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.538633 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ln7q6"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.571444 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lrp8\" (UniqueName: \"kubernetes.io/projected/d34acaa2-45bf-476e-8727-27b788e2887e-kube-api-access-4lrp8\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.571490 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-combined-ca-bundle\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.571551 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-operator-scripts\") pod \"barbican-ba6b-account-create-update-z9h5p\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.571604 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k49p\" (UniqueName: \"kubernetes.io/projected/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-kube-api-access-5k49p\") pod \"barbican-ba6b-account-create-update-z9h5p\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.571648 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59de480c-59b5-4a00-92c1-88c8fbc79366-operator-scripts\") pod \"cinder-db-create-mzvtm\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.571683 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbh4k\" (UniqueName: \"kubernetes.io/projected/59de480c-59b5-4a00-92c1-88c8fbc79366-kube-api-access-tbh4k\") pod \"cinder-db-create-mzvtm\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.571707 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-config-data\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.572187 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-operator-scripts\") pod \"barbican-ba6b-account-create-update-z9h5p\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.582601 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c547-account-create-update-4tbqv"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.583625 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.591718 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.596036 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c547-account-create-update-4tbqv"] Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.604443 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k49p\" (UniqueName: \"kubernetes.io/projected/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-kube-api-access-5k49p\") pod \"barbican-ba6b-account-create-update-z9h5p\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.658239 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.683653 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-combined-ca-bundle\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.683757 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-operator-scripts\") pod \"cinder-c547-account-create-update-4tbqv\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.683849 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kpkf\" (UniqueName: \"kubernetes.io/projected/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-kube-api-access-8kpkf\") pod \"cinder-c547-account-create-update-4tbqv\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.684066 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59de480c-59b5-4a00-92c1-88c8fbc79366-operator-scripts\") pod \"cinder-db-create-mzvtm\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.684160 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbh4k\" (UniqueName: \"kubernetes.io/projected/59de480c-59b5-4a00-92c1-88c8fbc79366-kube-api-access-tbh4k\") pod \"cinder-db-create-mzvtm\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.684225 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-config-data\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.684353 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lrp8\" (UniqueName: \"kubernetes.io/projected/d34acaa2-45bf-476e-8727-27b788e2887e-kube-api-access-4lrp8\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.691255 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59de480c-59b5-4a00-92c1-88c8fbc79366-operator-scripts\") pod \"cinder-db-create-mzvtm\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.715743 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-config-data\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.719435 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-combined-ca-bundle\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.733328 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbh4k\" (UniqueName: \"kubernetes.io/projected/59de480c-59b5-4a00-92c1-88c8fbc79366-kube-api-access-tbh4k\") pod \"cinder-db-create-mzvtm\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.768505 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lrp8\" (UniqueName: \"kubernetes.io/projected/d34acaa2-45bf-476e-8727-27b788e2887e-kube-api-access-4lrp8\") pod \"keystone-db-sync-ln7q6\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.783704 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.786369 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-operator-scripts\") pod \"cinder-c547-account-create-update-4tbqv\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.786489 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kpkf\" (UniqueName: \"kubernetes.io/projected/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-kube-api-access-8kpkf\") pod \"cinder-c547-account-create-update-4tbqv\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.787457 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-operator-scripts\") pod \"cinder-c547-account-create-update-4tbqv\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.822764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kpkf\" (UniqueName: \"kubernetes.io/projected/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-kube-api-access-8kpkf\") pod \"cinder-c547-account-create-update-4tbqv\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.866991 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:41 crc kubenswrapper[4750]: I0309 18:46:41.899218 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.084166 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.215845 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8k6sd" event={"ID":"46354cf1-acc8-45a7-a1e7-92f65230b565","Type":"ContainerStarted","Data":"1dff630734ab0900c47fda83ca71c84f6f2fee99fc3c331dc1ce4956c779f4e1"} Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.221757 4750 generic.go:334] "Generic (PLEG): container finished" podID="783ff672-9bd9-4092-8823-9a7bc9d2364a" containerID="f73e1c4aa1476494bae0883ca8680b278dc6dc9f17f0cb23e2ecc770967f6191" exitCode=0 Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.221854 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqfgs" event={"ID":"783ff672-9bd9-4092-8823-9a7bc9d2364a","Type":"ContainerDied","Data":"f73e1c4aa1476494bae0883ca8680b278dc6dc9f17f0cb23e2ecc770967f6191"} Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.254516 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8k6sd" podStartSLOduration=3.277402749 podStartE2EDuration="18.254489806s" podCreationTimestamp="2026-03-09 18:46:24 +0000 UTC" firstStartedPulling="2026-03-09 18:46:25.562081643 +0000 UTC m=+1266.904554041" lastFinishedPulling="2026-03-09 18:46:40.5391687 +0000 UTC m=+1281.881641098" observedRunningTime="2026-03-09 18:46:42.23728847 +0000 UTC m=+1283.579760878" watchObservedRunningTime="2026-03-09 18:46:42.254489806 +0000 UTC m=+1283.596962204" Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.258004 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"779ef3c76eee066cd25a806757da8086c496f5614a2e828d97ebf3ae3a275298"} Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.258068 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"ab29f6856785c236f628a9519150f83e031afa44ce0b034a739492a2669b000a"} Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.299136 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ba6b-account-create-update-z9h5p"] Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.766956 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rfqnd"] Mar 09 18:46:42 crc kubenswrapper[4750]: W0309 18:46:42.817808 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd34acaa2_45bf_476e_8727_27b788e2887e.slice/crio-1da551aa8ea989ecc6d31372f75ee7edbda8e00f538a59c8623372f0a1f937e9 WatchSource:0}: Error finding container 1da551aa8ea989ecc6d31372f75ee7edbda8e00f538a59c8623372f0a1f937e9: Status 404 returned error can't find the container with id 1da551aa8ea989ecc6d31372f75ee7edbda8e00f538a59c8623372f0a1f937e9 Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.825534 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ln7q6"] Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.839562 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mzvtm"] Mar 09 18:46:42 crc kubenswrapper[4750]: W0309 18:46:42.866321 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59de480c_59b5_4a00_92c1_88c8fbc79366.slice/crio-485fc13d25fb6f3c9cad9748521c0626eaf0f4c3858a464084135e41be539a3c WatchSource:0}: Error finding container 485fc13d25fb6f3c9cad9748521c0626eaf0f4c3858a464084135e41be539a3c: Status 404 returned error can't find the container with id 485fc13d25fb6f3c9cad9748521c0626eaf0f4c3858a464084135e41be539a3c Mar 09 18:46:42 crc kubenswrapper[4750]: I0309 18:46:42.933699 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c547-account-create-update-4tbqv"] Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.268172 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c547-account-create-update-4tbqv" event={"ID":"0e796a39-dd3b-4a32-ab3b-3c9e341381fb","Type":"ContainerStarted","Data":"d5cb3715c82cd76d7404f3419461e7fe7d6de1ca90cb57022beaec0781fa9ba6"} Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.274425 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"27406f345550e67745fa490a812eb4b299cc35abafbc928fa8acbc167a613aa1"} Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.275714 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rfqnd" event={"ID":"6c94fb97-754d-494d-ac4a-9fc9f3c9022c","Type":"ContainerStarted","Data":"32d0b6b208f088918ca551f8bf47df0d0f063b1d47c84d26d7381e999ea05d87"} Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.278013 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ln7q6" event={"ID":"d34acaa2-45bf-476e-8727-27b788e2887e","Type":"ContainerStarted","Data":"1da551aa8ea989ecc6d31372f75ee7edbda8e00f538a59c8623372f0a1f937e9"} Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.279534 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mzvtm" event={"ID":"59de480c-59b5-4a00-92c1-88c8fbc79366","Type":"ContainerStarted","Data":"485fc13d25fb6f3c9cad9748521c0626eaf0f4c3858a464084135e41be539a3c"} Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.282904 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ba6b-account-create-update-z9h5p" event={"ID":"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6","Type":"ContainerStarted","Data":"90277251ba6f9ada77e72495f221957c4824f976e512b472ce77961b9540d532"} Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.282944 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ba6b-account-create-update-z9h5p" event={"ID":"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6","Type":"ContainerStarted","Data":"aca7beb6863572952d7496f07de3216af2549e46131c4658f75653af6bcf516c"} Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.312225 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-ba6b-account-create-update-z9h5p" podStartSLOduration=2.312200236 podStartE2EDuration="2.312200236s" podCreationTimestamp="2026-03-09 18:46:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:43.31121073 +0000 UTC m=+1284.653683128" watchObservedRunningTime="2026-03-09 18:46:43.312200236 +0000 UTC m=+1284.654672634" Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.759839 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.824824 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/783ff672-9bd9-4092-8823-9a7bc9d2364a-operator-scripts\") pod \"783ff672-9bd9-4092-8823-9a7bc9d2364a\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.825078 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6nn2\" (UniqueName: \"kubernetes.io/projected/783ff672-9bd9-4092-8823-9a7bc9d2364a-kube-api-access-t6nn2\") pod \"783ff672-9bd9-4092-8823-9a7bc9d2364a\" (UID: \"783ff672-9bd9-4092-8823-9a7bc9d2364a\") " Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.827035 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/783ff672-9bd9-4092-8823-9a7bc9d2364a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "783ff672-9bd9-4092-8823-9a7bc9d2364a" (UID: "783ff672-9bd9-4092-8823-9a7bc9d2364a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.830505 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/783ff672-9bd9-4092-8823-9a7bc9d2364a-kube-api-access-t6nn2" (OuterVolumeSpecName: "kube-api-access-t6nn2") pod "783ff672-9bd9-4092-8823-9a7bc9d2364a" (UID: "783ff672-9bd9-4092-8823-9a7bc9d2364a"). InnerVolumeSpecName "kube-api-access-t6nn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.926992 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6nn2\" (UniqueName: \"kubernetes.io/projected/783ff672-9bd9-4092-8823-9a7bc9d2364a-kube-api-access-t6nn2\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:43 crc kubenswrapper[4750]: I0309 18:46:43.927041 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/783ff672-9bd9-4092-8823-9a7bc9d2364a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.303847 4750 generic.go:334] "Generic (PLEG): container finished" podID="59de480c-59b5-4a00-92c1-88c8fbc79366" containerID="bf9e1f72b2a0fcea54ff0ed44e2f2dc6919d164030ad2da3a7deb912556e0faa" exitCode=0 Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.303962 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mzvtm" event={"ID":"59de480c-59b5-4a00-92c1-88c8fbc79366","Type":"ContainerDied","Data":"bf9e1f72b2a0fcea54ff0ed44e2f2dc6919d164030ad2da3a7deb912556e0faa"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.320671 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-m55h9"] Mar 09 18:46:44 crc kubenswrapper[4750]: E0309 18:46:44.321282 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="783ff672-9bd9-4092-8823-9a7bc9d2364a" containerName="mariadb-account-create-update" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.321301 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="783ff672-9bd9-4092-8823-9a7bc9d2364a" containerName="mariadb-account-create-update" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.321493 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="783ff672-9bd9-4092-8823-9a7bc9d2364a" containerName="mariadb-account-create-update" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.322291 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.323560 4750 generic.go:334] "Generic (PLEG): container finished" podID="3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6" containerID="90277251ba6f9ada77e72495f221957c4824f976e512b472ce77961b9540d532" exitCode=0 Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.323658 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ba6b-account-create-update-z9h5p" event={"ID":"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6","Type":"ContainerDied","Data":"90277251ba6f9ada77e72495f221957c4824f976e512b472ce77961b9540d532"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.325494 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.327569 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-m55h9"] Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.349079 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-h7mwt" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.349126 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c547-account-create-update-4tbqv" event={"ID":"0e796a39-dd3b-4a32-ab3b-3c9e341381fb","Type":"ContainerStarted","Data":"6abb1c45bded4a1262afdeac6a5fe08053681bf937d7ae14cdcbfa48e057c5f9"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.398037 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"17442cd498c8e18ad1cad2d7a27aed8df35e94263fa361c196d1a21dd066e1c0"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.398098 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"bb57300fb659ddc723bacce39fcd9ab9e49491ec62b9304124bca4518df69bd0"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.412606 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6lf8\" (UniqueName: \"kubernetes.io/projected/8a84f925-3ae7-44b6-9e09-76da4848f588-kube-api-access-p6lf8\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.412813 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-config-data\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.412851 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-combined-ca-bundle\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.413050 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-db-sync-config-data\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.418101 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ca49786-3511-4719-90de-963bafdd8429","Type":"ContainerStarted","Data":"8cb5f7f9b23e82913430d0292707f63833fabf1fd7d6ec68d5e4e15b3cfe10f5"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.418152 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ca49786-3511-4719-90de-963bafdd8429","Type":"ContainerStarted","Data":"70392316fac60bc4ece60b3c8687f9d6bd0f2de3aad4709332e81f46db4f37db"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.438016 4750 generic.go:334] "Generic (PLEG): container finished" podID="6c94fb97-754d-494d-ac4a-9fc9f3c9022c" containerID="0c84c96140e622b0b2db5d836b37df651d1396b87c87bdee6df1263672f98fde" exitCode=0 Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.438128 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rfqnd" event={"ID":"6c94fb97-754d-494d-ac4a-9fc9f3c9022c","Type":"ContainerDied","Data":"0c84c96140e622b0b2db5d836b37df651d1396b87c87bdee6df1263672f98fde"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.442483 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqfgs" event={"ID":"783ff672-9bd9-4092-8823-9a7bc9d2364a","Type":"ContainerDied","Data":"fa85a72a45c57ebd0f46a9572e66eed4f7be97d77f0b461864664c48472bcbe6"} Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.442514 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa85a72a45c57ebd0f46a9572e66eed4f7be97d77f0b461864664c48472bcbe6" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.442601 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqfgs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.461989 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-c547-account-create-update-4tbqv" podStartSLOduration=3.461969043 podStartE2EDuration="3.461969043s" podCreationTimestamp="2026-03-09 18:46:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:44.46110983 +0000 UTC m=+1285.803582228" watchObservedRunningTime="2026-03-09 18:46:44.461969043 +0000 UTC m=+1285.804441441" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.498273 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-57gqs"] Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.499445 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.509061 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-57gqs"] Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.514466 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-combined-ca-bundle\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.514609 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-db-sync-config-data\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.514717 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6lf8\" (UniqueName: \"kubernetes.io/projected/8a84f925-3ae7-44b6-9e09-76da4848f588-kube-api-access-p6lf8\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.514806 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-config-data\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.523842 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-78c4-account-create-update-8v6f7"] Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.524883 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.530019 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.535555 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-db-sync-config-data\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.536984 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-config-data\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.540401 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=22.540374149 podStartE2EDuration="22.540374149s" podCreationTimestamp="2026-03-09 18:46:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:44.529895034 +0000 UTC m=+1285.872367432" watchObservedRunningTime="2026-03-09 18:46:44.540374149 +0000 UTC m=+1285.882846557" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.543201 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78c4-account-create-update-8v6f7"] Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.565933 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-combined-ca-bundle\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.571254 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6lf8\" (UniqueName: \"kubernetes.io/projected/8a84f925-3ae7-44b6-9e09-76da4848f588-kube-api-access-p6lf8\") pod \"watcher-db-sync-m55h9\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.616942 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfxwz\" (UniqueName: \"kubernetes.io/projected/6449a6c0-9e32-4b3c-b513-9672b6edb485-kube-api-access-jfxwz\") pod \"neutron-db-create-57gqs\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.617096 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6449a6c0-9e32-4b3c-b513-9672b6edb485-operator-scripts\") pod \"neutron-db-create-57gqs\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.689178 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-m55h9" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.718552 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f69cd0cb-75f0-4e75-825b-3f578b98f056-operator-scripts\") pod \"neutron-78c4-account-create-update-8v6f7\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.718679 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfxwz\" (UniqueName: \"kubernetes.io/projected/6449a6c0-9e32-4b3c-b513-9672b6edb485-kube-api-access-jfxwz\") pod \"neutron-db-create-57gqs\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.718735 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6449a6c0-9e32-4b3c-b513-9672b6edb485-operator-scripts\") pod \"neutron-db-create-57gqs\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.718765 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdfgg\" (UniqueName: \"kubernetes.io/projected/f69cd0cb-75f0-4e75-825b-3f578b98f056-kube-api-access-xdfgg\") pod \"neutron-78c4-account-create-update-8v6f7\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.719800 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6449a6c0-9e32-4b3c-b513-9672b6edb485-operator-scripts\") pod \"neutron-db-create-57gqs\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.752285 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfxwz\" (UniqueName: \"kubernetes.io/projected/6449a6c0-9e32-4b3c-b513-9672b6edb485-kube-api-access-jfxwz\") pod \"neutron-db-create-57gqs\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.820441 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdfgg\" (UniqueName: \"kubernetes.io/projected/f69cd0cb-75f0-4e75-825b-3f578b98f056-kube-api-access-xdfgg\") pod \"neutron-78c4-account-create-update-8v6f7\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.820521 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f69cd0cb-75f0-4e75-825b-3f578b98f056-operator-scripts\") pod \"neutron-78c4-account-create-update-8v6f7\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.821327 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f69cd0cb-75f0-4e75-825b-3f578b98f056-operator-scripts\") pod \"neutron-78c4-account-create-update-8v6f7\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.846200 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdfgg\" (UniqueName: \"kubernetes.io/projected/f69cd0cb-75f0-4e75-825b-3f578b98f056-kube-api-access-xdfgg\") pod \"neutron-78c4-account-create-update-8v6f7\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.939037 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:44 crc kubenswrapper[4750]: I0309 18:46:44.957456 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.328498 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-m55h9"] Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.481961 4750 generic.go:334] "Generic (PLEG): container finished" podID="0e796a39-dd3b-4a32-ab3b-3c9e341381fb" containerID="6abb1c45bded4a1262afdeac6a5fe08053681bf937d7ae14cdcbfa48e057c5f9" exitCode=0 Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.482051 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c547-account-create-update-4tbqv" event={"ID":"0e796a39-dd3b-4a32-ab3b-3c9e341381fb","Type":"ContainerDied","Data":"6abb1c45bded4a1262afdeac6a5fe08053681bf937d7ae14cdcbfa48e057c5f9"} Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.505228 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"76bc87fa0e6902df8b606f0e0f82ae6465bcdc57e22d66a67d332b00574908b7"} Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.505680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"1c896e587ad89f5f21ab0b3b794bb988d0db353d868140124a4f4a1a7b99e36f"} Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.505699 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"7ab9dd65a277de226a2ebfd456ccf9488aa925d81129136cece6c9eaf1311d5e"} Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.517829 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-m55h9" event={"ID":"8a84f925-3ae7-44b6-9e09-76da4848f588","Type":"ContainerStarted","Data":"8eaa212ae68f7cb68b551016930a6ef5f2a8757e747211d11e0d3d3a0c0363b4"} Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.637164 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-57gqs"] Mar 09 18:46:45 crc kubenswrapper[4750]: W0309 18:46:45.652001 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6449a6c0_9e32_4b3c_b513_9672b6edb485.slice/crio-46713e5c32256a3ce5ef026e01d84e80fe3eecee3e42f4a8185a713698b8877e WatchSource:0}: Error finding container 46713e5c32256a3ce5ef026e01d84e80fe3eecee3e42f4a8185a713698b8877e: Status 404 returned error can't find the container with id 46713e5c32256a3ce5ef026e01d84e80fe3eecee3e42f4a8185a713698b8877e Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.740451 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78c4-account-create-update-8v6f7"] Mar 09 18:46:45 crc kubenswrapper[4750]: W0309 18:46:45.750645 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf69cd0cb_75f0_4e75_825b_3f578b98f056.slice/crio-5f9732c6d8c1b3d4a01b502b35260f786771a937bf968853029082d1d227e44c WatchSource:0}: Error finding container 5f9732c6d8c1b3d4a01b502b35260f786771a937bf968853029082d1d227e44c: Status 404 returned error can't find the container with id 5f9732c6d8c1b3d4a01b502b35260f786771a937bf968853029082d1d227e44c Mar 09 18:46:45 crc kubenswrapper[4750]: I0309 18:46:45.959047 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.059336 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59de480c-59b5-4a00-92c1-88c8fbc79366-operator-scripts\") pod \"59de480c-59b5-4a00-92c1-88c8fbc79366\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.059549 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbh4k\" (UniqueName: \"kubernetes.io/projected/59de480c-59b5-4a00-92c1-88c8fbc79366-kube-api-access-tbh4k\") pod \"59de480c-59b5-4a00-92c1-88c8fbc79366\" (UID: \"59de480c-59b5-4a00-92c1-88c8fbc79366\") " Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.062070 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59de480c-59b5-4a00-92c1-88c8fbc79366-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "59de480c-59b5-4a00-92c1-88c8fbc79366" (UID: "59de480c-59b5-4a00-92c1-88c8fbc79366"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.063909 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.072491 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59de480c-59b5-4a00-92c1-88c8fbc79366-kube-api-access-tbh4k" (OuterVolumeSpecName: "kube-api-access-tbh4k") pod "59de480c-59b5-4a00-92c1-88c8fbc79366" (UID: "59de480c-59b5-4a00-92c1-88c8fbc79366"). InnerVolumeSpecName "kube-api-access-tbh4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.078547 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.165960 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gssq6\" (UniqueName: \"kubernetes.io/projected/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-kube-api-access-gssq6\") pod \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.166178 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-operator-scripts\") pod \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.166245 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k49p\" (UniqueName: \"kubernetes.io/projected/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-kube-api-access-5k49p\") pod \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\" (UID: \"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6\") " Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.166267 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-operator-scripts\") pod \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\" (UID: \"6c94fb97-754d-494d-ac4a-9fc9f3c9022c\") " Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.166649 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbh4k\" (UniqueName: \"kubernetes.io/projected/59de480c-59b5-4a00-92c1-88c8fbc79366-kube-api-access-tbh4k\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.166664 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59de480c-59b5-4a00-92c1-88c8fbc79366-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.167235 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c94fb97-754d-494d-ac4a-9fc9f3c9022c" (UID: "6c94fb97-754d-494d-ac4a-9fc9f3c9022c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.168630 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6" (UID: "3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.170428 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-kube-api-access-gssq6" (OuterVolumeSpecName: "kube-api-access-gssq6") pod "6c94fb97-754d-494d-ac4a-9fc9f3c9022c" (UID: "6c94fb97-754d-494d-ac4a-9fc9f3c9022c"). InnerVolumeSpecName "kube-api-access-gssq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.173029 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-kube-api-access-5k49p" (OuterVolumeSpecName: "kube-api-access-5k49p") pod "3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6" (UID: "3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6"). InnerVolumeSpecName "kube-api-access-5k49p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.268227 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.268272 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k49p\" (UniqueName: \"kubernetes.io/projected/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6-kube-api-access-5k49p\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.268289 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.268301 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gssq6\" (UniqueName: \"kubernetes.io/projected/6c94fb97-754d-494d-ac4a-9fc9f3c9022c-kube-api-access-gssq6\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.578060 4750 generic.go:334] "Generic (PLEG): container finished" podID="6449a6c0-9e32-4b3c-b513-9672b6edb485" containerID="0fb3a6126fd1cc367974868591fa5353b289803c04067a47eb4f58f16eb88a99" exitCode=0 Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.578156 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-57gqs" event={"ID":"6449a6c0-9e32-4b3c-b513-9672b6edb485","Type":"ContainerDied","Data":"0fb3a6126fd1cc367974868591fa5353b289803c04067a47eb4f58f16eb88a99"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.578192 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-57gqs" event={"ID":"6449a6c0-9e32-4b3c-b513-9672b6edb485","Type":"ContainerStarted","Data":"46713e5c32256a3ce5ef026e01d84e80fe3eecee3e42f4a8185a713698b8877e"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.580460 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ba6b-account-create-update-z9h5p" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.580708 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ba6b-account-create-update-z9h5p" event={"ID":"3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6","Type":"ContainerDied","Data":"aca7beb6863572952d7496f07de3216af2549e46131c4658f75653af6bcf516c"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.581093 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aca7beb6863572952d7496f07de3216af2549e46131c4658f75653af6bcf516c" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.604972 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"f29e70bf11ff40231ed55aee5c25eb4ff89e6abe2f2f21c7c5d7d2501168ebd9"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.605080 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cd12d247-7fd6-468d-8450-c395c4ee57ac","Type":"ContainerStarted","Data":"5d1995beac7fdfbd637a9bb722d413fea0fc3cbf18a6d52e5039c18c5ef88a0d"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.607908 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rfqnd" event={"ID":"6c94fb97-754d-494d-ac4a-9fc9f3c9022c","Type":"ContainerDied","Data":"32d0b6b208f088918ca551f8bf47df0d0f063b1d47c84d26d7381e999ea05d87"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.607970 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32d0b6b208f088918ca551f8bf47df0d0f063b1d47c84d26d7381e999ea05d87" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.607966 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rfqnd" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.612666 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mzvtm" event={"ID":"59de480c-59b5-4a00-92c1-88c8fbc79366","Type":"ContainerDied","Data":"485fc13d25fb6f3c9cad9748521c0626eaf0f4c3858a464084135e41be539a3c"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.612732 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="485fc13d25fb6f3c9cad9748521c0626eaf0f4c3858a464084135e41be539a3c" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.612693 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mzvtm" Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.614466 4750 generic.go:334] "Generic (PLEG): container finished" podID="f69cd0cb-75f0-4e75-825b-3f578b98f056" containerID="1f493884b398fc5c6bbca68a79dd5d8d92588c11220019ac57895b7872c5be67" exitCode=0 Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.614747 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78c4-account-create-update-8v6f7" event={"ID":"f69cd0cb-75f0-4e75-825b-3f578b98f056","Type":"ContainerDied","Data":"1f493884b398fc5c6bbca68a79dd5d8d92588c11220019ac57895b7872c5be67"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.614822 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78c4-account-create-update-8v6f7" event={"ID":"f69cd0cb-75f0-4e75-825b-3f578b98f056","Type":"ContainerStarted","Data":"5f9732c6d8c1b3d4a01b502b35260f786771a937bf968853029082d1d227e44c"} Mar 09 18:46:46 crc kubenswrapper[4750]: I0309 18:46:46.665700 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.798134575 podStartE2EDuration="50.665644658s" podCreationTimestamp="2026-03-09 18:45:56 +0000 UTC" firstStartedPulling="2026-03-09 18:46:30.597082454 +0000 UTC m=+1271.939554852" lastFinishedPulling="2026-03-09 18:46:43.464592537 +0000 UTC m=+1284.807064935" observedRunningTime="2026-03-09 18:46:46.647401842 +0000 UTC m=+1287.989874270" watchObservedRunningTime="2026-03-09 18:46:46.665644658 +0000 UTC m=+1288.008117076" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.023088 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d9fd9df5f-h7qcs"] Mar 09 18:46:47 crc kubenswrapper[4750]: E0309 18:46:47.023546 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c94fb97-754d-494d-ac4a-9fc9f3c9022c" containerName="mariadb-database-create" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.023563 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c94fb97-754d-494d-ac4a-9fc9f3c9022c" containerName="mariadb-database-create" Mar 09 18:46:47 crc kubenswrapper[4750]: E0309 18:46:47.023578 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6" containerName="mariadb-account-create-update" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.023585 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6" containerName="mariadb-account-create-update" Mar 09 18:46:47 crc kubenswrapper[4750]: E0309 18:46:47.023594 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59de480c-59b5-4a00-92c1-88c8fbc79366" containerName="mariadb-database-create" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.023600 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="59de480c-59b5-4a00-92c1-88c8fbc79366" containerName="mariadb-database-create" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.023805 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c94fb97-754d-494d-ac4a-9fc9f3c9022c" containerName="mariadb-database-create" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.023855 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="59de480c-59b5-4a00-92c1-88c8fbc79366" containerName="mariadb-database-create" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.023874 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6" containerName="mariadb-account-create-update" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.025616 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.037806 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.057806 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d9fd9df5f-h7qcs"] Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.093045 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-config\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.093114 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.093176 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-svc\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.093249 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-swift-storage-0\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.093277 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-sb\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.093308 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnwt7\" (UniqueName: \"kubernetes.io/projected/5b794abe-5fa4-47bd-bc10-6a871c5ee196-kube-api-access-tnwt7\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.194618 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-swift-storage-0\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.194698 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-sb\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.194732 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnwt7\" (UniqueName: \"kubernetes.io/projected/5b794abe-5fa4-47bd-bc10-6a871c5ee196-kube-api-access-tnwt7\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.194761 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-config\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.194787 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.194833 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-svc\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.196062 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-svc\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.196110 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-swift-storage-0\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.196594 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-sb\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.196700 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-config\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.196787 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.222962 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnwt7\" (UniqueName: \"kubernetes.io/projected/5b794abe-5fa4-47bd-bc10-6a871c5ee196-kube-api-access-tnwt7\") pod \"dnsmasq-dns-6d9fd9df5f-h7qcs\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:47 crc kubenswrapper[4750]: I0309 18:46:47.369439 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:48 crc kubenswrapper[4750]: I0309 18:46:48.262317 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:50 crc kubenswrapper[4750]: I0309 18:46:50.827594 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:50 crc kubenswrapper[4750]: I0309 18:46:50.866450 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfxwz\" (UniqueName: \"kubernetes.io/projected/6449a6c0-9e32-4b3c-b513-9672b6edb485-kube-api-access-jfxwz\") pod \"6449a6c0-9e32-4b3c-b513-9672b6edb485\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " Mar 09 18:46:50 crc kubenswrapper[4750]: I0309 18:46:50.866599 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6449a6c0-9e32-4b3c-b513-9672b6edb485-operator-scripts\") pod \"6449a6c0-9e32-4b3c-b513-9672b6edb485\" (UID: \"6449a6c0-9e32-4b3c-b513-9672b6edb485\") " Mar 09 18:46:50 crc kubenswrapper[4750]: I0309 18:46:50.868364 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6449a6c0-9e32-4b3c-b513-9672b6edb485-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6449a6c0-9e32-4b3c-b513-9672b6edb485" (UID: "6449a6c0-9e32-4b3c-b513-9672b6edb485"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:50 crc kubenswrapper[4750]: I0309 18:46:50.874099 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6449a6c0-9e32-4b3c-b513-9672b6edb485-kube-api-access-jfxwz" (OuterVolumeSpecName: "kube-api-access-jfxwz") pod "6449a6c0-9e32-4b3c-b513-9672b6edb485" (UID: "6449a6c0-9e32-4b3c-b513-9672b6edb485"). InnerVolumeSpecName "kube-api-access-jfxwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:50 crc kubenswrapper[4750]: I0309 18:46:50.968576 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfxwz\" (UniqueName: \"kubernetes.io/projected/6449a6c0-9e32-4b3c-b513-9672b6edb485-kube-api-access-jfxwz\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:50 crc kubenswrapper[4750]: I0309 18:46:50.968620 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6449a6c0-9e32-4b3c-b513-9672b6edb485-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:51 crc kubenswrapper[4750]: I0309 18:46:51.682463 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-57gqs" event={"ID":"6449a6c0-9e32-4b3c-b513-9672b6edb485","Type":"ContainerDied","Data":"46713e5c32256a3ce5ef026e01d84e80fe3eecee3e42f4a8185a713698b8877e"} Mar 09 18:46:51 crc kubenswrapper[4750]: I0309 18:46:51.682820 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46713e5c32256a3ce5ef026e01d84e80fe3eecee3e42f4a8185a713698b8877e" Mar 09 18:46:51 crc kubenswrapper[4750]: I0309 18:46:51.682659 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-57gqs" Mar 09 18:46:51 crc kubenswrapper[4750]: I0309 18:46:51.743526 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:46:51 crc kubenswrapper[4750]: I0309 18:46:51.743587 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:46:52 crc kubenswrapper[4750]: I0309 18:46:52.696791 4750 generic.go:334] "Generic (PLEG): container finished" podID="46354cf1-acc8-45a7-a1e7-92f65230b565" containerID="1dff630734ab0900c47fda83ca71c84f6f2fee99fc3c331dc1ce4956c779f4e1" exitCode=0 Mar 09 18:46:52 crc kubenswrapper[4750]: I0309 18:46:52.696882 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8k6sd" event={"ID":"46354cf1-acc8-45a7-a1e7-92f65230b565","Type":"ContainerDied","Data":"1dff630734ab0900c47fda83ca71c84f6f2fee99fc3c331dc1ce4956c779f4e1"} Mar 09 18:46:53 crc kubenswrapper[4750]: I0309 18:46:53.262336 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:53 crc kubenswrapper[4750]: I0309 18:46:53.281118 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:53 crc kubenswrapper[4750]: I0309 18:46:53.711592 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.865676 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.876328 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.892029 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998101 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfblk\" (UniqueName: \"kubernetes.io/projected/46354cf1-acc8-45a7-a1e7-92f65230b565-kube-api-access-cfblk\") pod \"46354cf1-acc8-45a7-a1e7-92f65230b565\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998191 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-combined-ca-bundle\") pod \"46354cf1-acc8-45a7-a1e7-92f65230b565\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998230 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f69cd0cb-75f0-4e75-825b-3f578b98f056-operator-scripts\") pod \"f69cd0cb-75f0-4e75-825b-3f578b98f056\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998281 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kpkf\" (UniqueName: \"kubernetes.io/projected/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-kube-api-access-8kpkf\") pod \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998356 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdfgg\" (UniqueName: \"kubernetes.io/projected/f69cd0cb-75f0-4e75-825b-3f578b98f056-kube-api-access-xdfgg\") pod \"f69cd0cb-75f0-4e75-825b-3f578b98f056\" (UID: \"f69cd0cb-75f0-4e75-825b-3f578b98f056\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998497 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-config-data\") pod \"46354cf1-acc8-45a7-a1e7-92f65230b565\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998521 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-db-sync-config-data\") pod \"46354cf1-acc8-45a7-a1e7-92f65230b565\" (UID: \"46354cf1-acc8-45a7-a1e7-92f65230b565\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.998550 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-operator-scripts\") pod \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\" (UID: \"0e796a39-dd3b-4a32-ab3b-3c9e341381fb\") " Mar 09 18:46:54 crc kubenswrapper[4750]: I0309 18:46:54.999096 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f69cd0cb-75f0-4e75-825b-3f578b98f056-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f69cd0cb-75f0-4e75-825b-3f578b98f056" (UID: "f69cd0cb-75f0-4e75-825b-3f578b98f056"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:54.999908 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e796a39-dd3b-4a32-ab3b-3c9e341381fb" (UID: "0e796a39-dd3b-4a32-ab3b-3c9e341381fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.005722 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "46354cf1-acc8-45a7-a1e7-92f65230b565" (UID: "46354cf1-acc8-45a7-a1e7-92f65230b565"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.005934 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-kube-api-access-8kpkf" (OuterVolumeSpecName: "kube-api-access-8kpkf") pod "0e796a39-dd3b-4a32-ab3b-3c9e341381fb" (UID: "0e796a39-dd3b-4a32-ab3b-3c9e341381fb"). InnerVolumeSpecName "kube-api-access-8kpkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.007918 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46354cf1-acc8-45a7-a1e7-92f65230b565-kube-api-access-cfblk" (OuterVolumeSpecName: "kube-api-access-cfblk") pod "46354cf1-acc8-45a7-a1e7-92f65230b565" (UID: "46354cf1-acc8-45a7-a1e7-92f65230b565"). InnerVolumeSpecName "kube-api-access-cfblk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.020229 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69cd0cb-75f0-4e75-825b-3f578b98f056-kube-api-access-xdfgg" (OuterVolumeSpecName: "kube-api-access-xdfgg") pod "f69cd0cb-75f0-4e75-825b-3f578b98f056" (UID: "f69cd0cb-75f0-4e75-825b-3f578b98f056"). InnerVolumeSpecName "kube-api-access-xdfgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.030435 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46354cf1-acc8-45a7-a1e7-92f65230b565" (UID: "46354cf1-acc8-45a7-a1e7-92f65230b565"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.046844 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-config-data" (OuterVolumeSpecName: "config-data") pod "46354cf1-acc8-45a7-a1e7-92f65230b565" (UID: "46354cf1-acc8-45a7-a1e7-92f65230b565"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101791 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101829 4750 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101844 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101856 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfblk\" (UniqueName: \"kubernetes.io/projected/46354cf1-acc8-45a7-a1e7-92f65230b565-kube-api-access-cfblk\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101869 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46354cf1-acc8-45a7-a1e7-92f65230b565-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101881 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f69cd0cb-75f0-4e75-825b-3f578b98f056-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101892 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kpkf\" (UniqueName: \"kubernetes.io/projected/0e796a39-dd3b-4a32-ab3b-3c9e341381fb-kube-api-access-8kpkf\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.101903 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdfgg\" (UniqueName: \"kubernetes.io/projected/f69cd0cb-75f0-4e75-825b-3f578b98f056-kube-api-access-xdfgg\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.726290 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c547-account-create-update-4tbqv" event={"ID":"0e796a39-dd3b-4a32-ab3b-3c9e341381fb","Type":"ContainerDied","Data":"d5cb3715c82cd76d7404f3419461e7fe7d6de1ca90cb57022beaec0781fa9ba6"} Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.726524 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5cb3715c82cd76d7404f3419461e7fe7d6de1ca90cb57022beaec0781fa9ba6" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.726592 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c547-account-create-update-4tbqv" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.731023 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-m55h9" event={"ID":"8a84f925-3ae7-44b6-9e09-76da4848f588","Type":"ContainerStarted","Data":"c1c097139ef200f5474967dd9000e7eb1664f09c25fc16f9adfbce7a95aae746"} Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.732568 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8k6sd" event={"ID":"46354cf1-acc8-45a7-a1e7-92f65230b565","Type":"ContainerDied","Data":"b5c483538d89322d5430e91bbbb6463c20bcd8882640341ef75402a997dea45e"} Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.732591 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5c483538d89322d5430e91bbbb6463c20bcd8882640341ef75402a997dea45e" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.732695 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8k6sd" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.733953 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ln7q6" event={"ID":"d34acaa2-45bf-476e-8727-27b788e2887e","Type":"ContainerStarted","Data":"43104fa1b64d7ee48e967155b623a9515efab604fcddbaddbd8f401c10b0bf98"} Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.736543 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78c4-account-create-update-8v6f7" event={"ID":"f69cd0cb-75f0-4e75-825b-3f578b98f056","Type":"ContainerDied","Data":"5f9732c6d8c1b3d4a01b502b35260f786771a937bf968853029082d1d227e44c"} Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.736576 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f9732c6d8c1b3d4a01b502b35260f786771a937bf968853029082d1d227e44c" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.736653 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78c4-account-create-update-8v6f7" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.776862 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-m55h9" podStartSLOduration=1.671646323 podStartE2EDuration="11.776843875s" podCreationTimestamp="2026-03-09 18:46:44 +0000 UTC" firstStartedPulling="2026-03-09 18:46:45.341873454 +0000 UTC m=+1286.684345852" lastFinishedPulling="2026-03-09 18:46:55.447070966 +0000 UTC m=+1296.789543404" observedRunningTime="2026-03-09 18:46:55.772457796 +0000 UTC m=+1297.114930214" watchObservedRunningTime="2026-03-09 18:46:55.776843875 +0000 UTC m=+1297.119316273" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.790996 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ln7q6" podStartSLOduration=2.2084616280000002 podStartE2EDuration="14.790976159s" podCreationTimestamp="2026-03-09 18:46:41 +0000 UTC" firstStartedPulling="2026-03-09 18:46:42.821960298 +0000 UTC m=+1284.164432696" lastFinishedPulling="2026-03-09 18:46:55.404474829 +0000 UTC m=+1296.746947227" observedRunningTime="2026-03-09 18:46:55.789980712 +0000 UTC m=+1297.132453190" watchObservedRunningTime="2026-03-09 18:46:55.790976159 +0000 UTC m=+1297.133448557" Mar 09 18:46:55 crc kubenswrapper[4750]: I0309 18:46:55.855970 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d9fd9df5f-h7qcs"] Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.444520 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9fd9df5f-h7qcs"] Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.482376 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f8c47bdf-hmx4v"] Mar 09 18:46:56 crc kubenswrapper[4750]: E0309 18:46:56.482923 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e796a39-dd3b-4a32-ab3b-3c9e341381fb" containerName="mariadb-account-create-update" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.482985 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e796a39-dd3b-4a32-ab3b-3c9e341381fb" containerName="mariadb-account-create-update" Mar 09 18:46:56 crc kubenswrapper[4750]: E0309 18:46:56.483046 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46354cf1-acc8-45a7-a1e7-92f65230b565" containerName="glance-db-sync" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.483119 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="46354cf1-acc8-45a7-a1e7-92f65230b565" containerName="glance-db-sync" Mar 09 18:46:56 crc kubenswrapper[4750]: E0309 18:46:56.483177 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6449a6c0-9e32-4b3c-b513-9672b6edb485" containerName="mariadb-database-create" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.483222 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6449a6c0-9e32-4b3c-b513-9672b6edb485" containerName="mariadb-database-create" Mar 09 18:46:56 crc kubenswrapper[4750]: E0309 18:46:56.483280 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69cd0cb-75f0-4e75-825b-3f578b98f056" containerName="mariadb-account-create-update" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.483324 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69cd0cb-75f0-4e75-825b-3f578b98f056" containerName="mariadb-account-create-update" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.483503 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e796a39-dd3b-4a32-ab3b-3c9e341381fb" containerName="mariadb-account-create-update" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.483576 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6449a6c0-9e32-4b3c-b513-9672b6edb485" containerName="mariadb-database-create" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.483654 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="46354cf1-acc8-45a7-a1e7-92f65230b565" containerName="glance-db-sync" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.483726 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69cd0cb-75f0-4e75-825b-3f578b98f056" containerName="mariadb-account-create-update" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.484674 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.502783 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f8c47bdf-hmx4v"] Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.528590 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.528696 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.528724 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9gk9\" (UniqueName: \"kubernetes.io/projected/5772d728-1aa9-4b36-925a-d6d7694de34c-kube-api-access-t9gk9\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.528747 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-svc\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.528776 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.529158 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-config\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.630657 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.631007 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9gk9\" (UniqueName: \"kubernetes.io/projected/5772d728-1aa9-4b36-925a-d6d7694de34c-kube-api-access-t9gk9\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.631111 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-svc\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.631198 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.631307 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-config\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.631443 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.631564 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-swift-storage-0\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.631927 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-nb\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.632225 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-svc\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.632412 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-sb\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.632692 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-config\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.652178 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9gk9\" (UniqueName: \"kubernetes.io/projected/5772d728-1aa9-4b36-925a-d6d7694de34c-kube-api-access-t9gk9\") pod \"dnsmasq-dns-85f8c47bdf-hmx4v\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.760214 4750 generic.go:334] "Generic (PLEG): container finished" podID="5b794abe-5fa4-47bd-bc10-6a871c5ee196" containerID="09e6dc0177793a6e0981ce19c7445c8f4f55867b8d530ec9b1f0992c3c3b51fe" exitCode=0 Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.760467 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" event={"ID":"5b794abe-5fa4-47bd-bc10-6a871c5ee196","Type":"ContainerDied","Data":"09e6dc0177793a6e0981ce19c7445c8f4f55867b8d530ec9b1f0992c3c3b51fe"} Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.760525 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" event={"ID":"5b794abe-5fa4-47bd-bc10-6a871c5ee196","Type":"ContainerStarted","Data":"98b549ae18c844c82dc1c7b20560aff1d6fa2ce82a635dc7efc5587c30cd4dac"} Mar 09 18:46:56 crc kubenswrapper[4750]: I0309 18:46:56.812962 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:57 crc kubenswrapper[4750]: E0309 18:46:57.053401 4750 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Mar 09 18:46:57 crc kubenswrapper[4750]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/5b794abe-5fa4-47bd-bc10-6a871c5ee196/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 09 18:46:57 crc kubenswrapper[4750]: > podSandboxID="98b549ae18c844c82dc1c7b20560aff1d6fa2ce82a635dc7efc5587c30cd4dac" Mar 09 18:46:57 crc kubenswrapper[4750]: E0309 18:46:57.053997 4750 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 18:46:57 crc kubenswrapper[4750]: container &Container{Name:dnsmasq-dns,Image:38.102.83.110:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncbh5fh58h56bh64bh684h98h654h64bh687h645h54h548h87h59dh56dh655hd9hbfh87h5c9h68bh645h64h8bh8bh585h5bdh55ch546hfbhbdq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tnwt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6d9fd9df5f-h7qcs_openstack(5b794abe-5fa4-47bd-bc10-6a871c5ee196): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/5b794abe-5fa4-47bd-bc10-6a871c5ee196/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 09 18:46:57 crc kubenswrapper[4750]: > logger="UnhandledError" Mar 09 18:46:57 crc kubenswrapper[4750]: E0309 18:46:57.055727 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/5b794abe-5fa4-47bd-bc10-6a871c5ee196/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" podUID="5b794abe-5fa4-47bd-bc10-6a871c5ee196" Mar 09 18:46:57 crc kubenswrapper[4750]: I0309 18:46:57.315209 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f8c47bdf-hmx4v"] Mar 09 18:46:57 crc kubenswrapper[4750]: W0309 18:46:57.315918 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5772d728_1aa9_4b36_925a_d6d7694de34c.slice/crio-2d8be5663a529e390f7c12be785803d7e253d5195777096ab8a780da17ee00da WatchSource:0}: Error finding container 2d8be5663a529e390f7c12be785803d7e253d5195777096ab8a780da17ee00da: Status 404 returned error can't find the container with id 2d8be5663a529e390f7c12be785803d7e253d5195777096ab8a780da17ee00da Mar 09 18:46:57 crc kubenswrapper[4750]: I0309 18:46:57.774295 4750 generic.go:334] "Generic (PLEG): container finished" podID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerID="154ff451a12cb04e1f3e6673e5f88d9b1989498db68b23f07de84514fbf4a063" exitCode=0 Mar 09 18:46:57 crc kubenswrapper[4750]: I0309 18:46:57.774846 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" event={"ID":"5772d728-1aa9-4b36-925a-d6d7694de34c","Type":"ContainerDied","Data":"154ff451a12cb04e1f3e6673e5f88d9b1989498db68b23f07de84514fbf4a063"} Mar 09 18:46:57 crc kubenswrapper[4750]: I0309 18:46:57.774918 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" event={"ID":"5772d728-1aa9-4b36-925a-d6d7694de34c","Type":"ContainerStarted","Data":"2d8be5663a529e390f7c12be785803d7e253d5195777096ab8a780da17ee00da"} Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.207033 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.282241 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-nb\") pod \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.282323 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnwt7\" (UniqueName: \"kubernetes.io/projected/5b794abe-5fa4-47bd-bc10-6a871c5ee196-kube-api-access-tnwt7\") pod \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.282354 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-config\") pod \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.282461 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-sb\") pod \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.282520 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-svc\") pod \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.282548 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-swift-storage-0\") pod \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\" (UID: \"5b794abe-5fa4-47bd-bc10-6a871c5ee196\") " Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.326834 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b794abe-5fa4-47bd-bc10-6a871c5ee196-kube-api-access-tnwt7" (OuterVolumeSpecName: "kube-api-access-tnwt7") pod "5b794abe-5fa4-47bd-bc10-6a871c5ee196" (UID: "5b794abe-5fa4-47bd-bc10-6a871c5ee196"). InnerVolumeSpecName "kube-api-access-tnwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.355973 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-config" (OuterVolumeSpecName: "config") pod "5b794abe-5fa4-47bd-bc10-6a871c5ee196" (UID: "5b794abe-5fa4-47bd-bc10-6a871c5ee196"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.384130 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.384168 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnwt7\" (UniqueName: \"kubernetes.io/projected/5b794abe-5fa4-47bd-bc10-6a871c5ee196-kube-api-access-tnwt7\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.395671 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5b794abe-5fa4-47bd-bc10-6a871c5ee196" (UID: "5b794abe-5fa4-47bd-bc10-6a871c5ee196"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.396957 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b794abe-5fa4-47bd-bc10-6a871c5ee196" (UID: "5b794abe-5fa4-47bd-bc10-6a871c5ee196"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.403576 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5b794abe-5fa4-47bd-bc10-6a871c5ee196" (UID: "5b794abe-5fa4-47bd-bc10-6a871c5ee196"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.404161 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b794abe-5fa4-47bd-bc10-6a871c5ee196" (UID: "5b794abe-5fa4-47bd-bc10-6a871c5ee196"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.486288 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.486328 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.486337 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.486345 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b794abe-5fa4-47bd-bc10-6a871c5ee196-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.784961 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" event={"ID":"5772d728-1aa9-4b36-925a-d6d7694de34c","Type":"ContainerStarted","Data":"415d643fb68a9db0f410487650f656c384fd17e3e3e19205e7db418562a7617a"} Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.785084 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.788028 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" event={"ID":"5b794abe-5fa4-47bd-bc10-6a871c5ee196","Type":"ContainerDied","Data":"98b549ae18c844c82dc1c7b20560aff1d6fa2ce82a635dc7efc5587c30cd4dac"} Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.788094 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9fd9df5f-h7qcs" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.788106 4750 scope.go:117] "RemoveContainer" containerID="09e6dc0177793a6e0981ce19c7445c8f4f55867b8d530ec9b1f0992c3c3b51fe" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.812576 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" podStartSLOduration=2.8125540300000003 podStartE2EDuration="2.81255403s" podCreationTimestamp="2026-03-09 18:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:46:58.81145863 +0000 UTC m=+1300.153931038" watchObservedRunningTime="2026-03-09 18:46:58.81255403 +0000 UTC m=+1300.155026428" Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.925868 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9fd9df5f-h7qcs"] Mar 09 18:46:58 crc kubenswrapper[4750]: I0309 18:46:58.934183 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d9fd9df5f-h7qcs"] Mar 09 18:46:59 crc kubenswrapper[4750]: E0309 18:46:59.073854 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b794abe_5fa4_47bd_bc10_6a871c5ee196.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b794abe_5fa4_47bd_bc10_6a871c5ee196.slice/crio-98b549ae18c844c82dc1c7b20560aff1d6fa2ce82a635dc7efc5587c30cd4dac\": RecentStats: unable to find data in memory cache]" Mar 09 18:46:59 crc kubenswrapper[4750]: I0309 18:46:59.385090 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b794abe-5fa4-47bd-bc10-6a871c5ee196" path="/var/lib/kubelet/pods/5b794abe-5fa4-47bd-bc10-6a871c5ee196/volumes" Mar 09 18:47:00 crc kubenswrapper[4750]: I0309 18:47:00.833744 4750 generic.go:334] "Generic (PLEG): container finished" podID="d34acaa2-45bf-476e-8727-27b788e2887e" containerID="43104fa1b64d7ee48e967155b623a9515efab604fcddbaddbd8f401c10b0bf98" exitCode=0 Mar 09 18:47:00 crc kubenswrapper[4750]: I0309 18:47:00.833833 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ln7q6" event={"ID":"d34acaa2-45bf-476e-8727-27b788e2887e","Type":"ContainerDied","Data":"43104fa1b64d7ee48e967155b623a9515efab604fcddbaddbd8f401c10b0bf98"} Mar 09 18:47:00 crc kubenswrapper[4750]: I0309 18:47:00.836177 4750 generic.go:334] "Generic (PLEG): container finished" podID="8a84f925-3ae7-44b6-9e09-76da4848f588" containerID="c1c097139ef200f5474967dd9000e7eb1664f09c25fc16f9adfbce7a95aae746" exitCode=0 Mar 09 18:47:00 crc kubenswrapper[4750]: I0309 18:47:00.836224 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-m55h9" event={"ID":"8a84f925-3ae7-44b6-9e09-76da4848f588","Type":"ContainerDied","Data":"c1c097139ef200f5474967dd9000e7eb1664f09c25fc16f9adfbce7a95aae746"} Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.307689 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-m55h9" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.313237 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.361694 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-db-sync-config-data\") pod \"8a84f925-3ae7-44b6-9e09-76da4848f588\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.361891 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-combined-ca-bundle\") pod \"d34acaa2-45bf-476e-8727-27b788e2887e\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.361921 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-config-data\") pod \"d34acaa2-45bf-476e-8727-27b788e2887e\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.361987 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-combined-ca-bundle\") pod \"8a84f925-3ae7-44b6-9e09-76da4848f588\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.362028 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6lf8\" (UniqueName: \"kubernetes.io/projected/8a84f925-3ae7-44b6-9e09-76da4848f588-kube-api-access-p6lf8\") pod \"8a84f925-3ae7-44b6-9e09-76da4848f588\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.364599 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lrp8\" (UniqueName: \"kubernetes.io/projected/d34acaa2-45bf-476e-8727-27b788e2887e-kube-api-access-4lrp8\") pod \"d34acaa2-45bf-476e-8727-27b788e2887e\" (UID: \"d34acaa2-45bf-476e-8727-27b788e2887e\") " Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.364830 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-config-data\") pod \"8a84f925-3ae7-44b6-9e09-76da4848f588\" (UID: \"8a84f925-3ae7-44b6-9e09-76da4848f588\") " Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.377208 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8a84f925-3ae7-44b6-9e09-76da4848f588" (UID: "8a84f925-3ae7-44b6-9e09-76da4848f588"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.384057 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d34acaa2-45bf-476e-8727-27b788e2887e-kube-api-access-4lrp8" (OuterVolumeSpecName: "kube-api-access-4lrp8") pod "d34acaa2-45bf-476e-8727-27b788e2887e" (UID: "d34acaa2-45bf-476e-8727-27b788e2887e"). InnerVolumeSpecName "kube-api-access-4lrp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.393022 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a84f925-3ae7-44b6-9e09-76da4848f588-kube-api-access-p6lf8" (OuterVolumeSpecName: "kube-api-access-p6lf8") pod "8a84f925-3ae7-44b6-9e09-76da4848f588" (UID: "8a84f925-3ae7-44b6-9e09-76da4848f588"). InnerVolumeSpecName "kube-api-access-p6lf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.400477 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a84f925-3ae7-44b6-9e09-76da4848f588" (UID: "8a84f925-3ae7-44b6-9e09-76da4848f588"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.407295 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d34acaa2-45bf-476e-8727-27b788e2887e" (UID: "d34acaa2-45bf-476e-8727-27b788e2887e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.423768 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-config-data" (OuterVolumeSpecName: "config-data") pod "d34acaa2-45bf-476e-8727-27b788e2887e" (UID: "d34acaa2-45bf-476e-8727-27b788e2887e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.429786 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-config-data" (OuterVolumeSpecName: "config-data") pod "8a84f925-3ae7-44b6-9e09-76da4848f588" (UID: "8a84f925-3ae7-44b6-9e09-76da4848f588"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.467646 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.467679 4750 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.467700 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.467713 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34acaa2-45bf-476e-8727-27b788e2887e-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.467730 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a84f925-3ae7-44b6-9e09-76da4848f588-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.467741 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6lf8\" (UniqueName: \"kubernetes.io/projected/8a84f925-3ae7-44b6-9e09-76da4848f588-kube-api-access-p6lf8\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.467754 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lrp8\" (UniqueName: \"kubernetes.io/projected/d34acaa2-45bf-476e-8727-27b788e2887e-kube-api-access-4lrp8\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.856381 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ln7q6" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.856393 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ln7q6" event={"ID":"d34acaa2-45bf-476e-8727-27b788e2887e","Type":"ContainerDied","Data":"1da551aa8ea989ecc6d31372f75ee7edbda8e00f538a59c8623372f0a1f937e9"} Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.856445 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da551aa8ea989ecc6d31372f75ee7edbda8e00f538a59c8623372f0a1f937e9" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.858206 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-m55h9" event={"ID":"8a84f925-3ae7-44b6-9e09-76da4848f588","Type":"ContainerDied","Data":"8eaa212ae68f7cb68b551016930a6ef5f2a8757e747211d11e0d3d3a0c0363b4"} Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.858262 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eaa212ae68f7cb68b551016930a6ef5f2a8757e747211d11e0d3d3a0c0363b4" Mar 09 18:47:02 crc kubenswrapper[4750]: I0309 18:47:02.858331 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-m55h9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.160217 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f8c47bdf-hmx4v"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.160820 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" podUID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerName="dnsmasq-dns" containerID="cri-o://415d643fb68a9db0f410487650f656c384fd17e3e3e19205e7db418562a7617a" gracePeriod=10 Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.165586 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.207807 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6kmdp"] Mar 09 18:47:03 crc kubenswrapper[4750]: E0309 18:47:03.208172 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d34acaa2-45bf-476e-8727-27b788e2887e" containerName="keystone-db-sync" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.208190 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d34acaa2-45bf-476e-8727-27b788e2887e" containerName="keystone-db-sync" Mar 09 18:47:03 crc kubenswrapper[4750]: E0309 18:47:03.208239 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b794abe-5fa4-47bd-bc10-6a871c5ee196" containerName="init" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.208250 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b794abe-5fa4-47bd-bc10-6a871c5ee196" containerName="init" Mar 09 18:47:03 crc kubenswrapper[4750]: E0309 18:47:03.208267 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a84f925-3ae7-44b6-9e09-76da4848f588" containerName="watcher-db-sync" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.208274 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a84f925-3ae7-44b6-9e09-76da4848f588" containerName="watcher-db-sync" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.208434 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b794abe-5fa4-47bd-bc10-6a871c5ee196" containerName="init" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.208458 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a84f925-3ae7-44b6-9e09-76da4848f588" containerName="watcher-db-sync" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.208476 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d34acaa2-45bf-476e-8727-27b788e2887e" containerName="keystone-db-sync" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.209091 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.225280 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.225598 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.225788 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.225971 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6kmdp"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.229078 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9flmk" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.232543 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.251771 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89cbc8c6f-jpdk9"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.253709 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.291938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-combined-ca-bundle\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.292097 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-config-data\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.292313 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-fernet-keys\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.292409 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh89x\" (UniqueName: \"kubernetes.io/projected/60a57822-e425-4ef7-8da8-f83a0c0abe60-kube-api-access-sh89x\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.292448 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-credential-keys\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.292487 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-scripts\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424100 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-credential-keys\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424159 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-scripts\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424228 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-sb\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424252 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-config\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424271 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-combined-ca-bundle\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424310 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-svc\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424336 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-config-data\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424360 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-swift-storage-0\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424414 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-nb\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424456 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-fernet-keys\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424484 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nm5k\" (UniqueName: \"kubernetes.io/projected/3f1158dc-0354-469d-82f7-1d5e1a9de48f-kube-api-access-9nm5k\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.424517 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh89x\" (UniqueName: \"kubernetes.io/projected/60a57822-e425-4ef7-8da8-f83a0c0abe60-kube-api-access-sh89x\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.431751 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89cbc8c6f-jpdk9"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.445220 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-credential-keys\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.450985 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-scripts\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.458320 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-combined-ca-bundle\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.463423 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-fernet-keys\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.470509 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-config-data\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.498222 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh89x\" (UniqueName: \"kubernetes.io/projected/60a57822-e425-4ef7-8da8-f83a0c0abe60-kube-api-access-sh89x\") pod \"keystone-bootstrap-6kmdp\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.526482 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-svc\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.526582 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-swift-storage-0\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.528001 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-nb\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.528087 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nm5k\" (UniqueName: \"kubernetes.io/projected/3f1158dc-0354-469d-82f7-1d5e1a9de48f-kube-api-access-9nm5k\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.528181 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-sb\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.528205 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-config\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.529308 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-config\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.529872 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-svc\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.536079 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-swift-storage-0\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.537556 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-nb\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.538083 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-sb\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.574393 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nm5k\" (UniqueName: \"kubernetes.io/projected/3f1158dc-0354-469d-82f7-1d5e1a9de48f-kube-api-access-9nm5k\") pod \"dnsmasq-dns-89cbc8c6f-jpdk9\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.608961 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.611142 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.617434 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.627549 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-h7mwt" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.634498 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.636179 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.639595 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.653054 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.664848 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.690604 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.691755 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.735529 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-config-data\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736162 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh7tw\" (UniqueName: \"kubernetes.io/projected/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-kube-api-access-lh7tw\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736281 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d32a26f-bd14-41a7-872b-e102ea95f696-logs\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736332 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lppbg\" (UniqueName: \"kubernetes.io/projected/9d32a26f-bd14-41a7-872b-e102ea95f696-kube-api-access-lppbg\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736386 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-config-data\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736426 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-logs\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736447 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736475 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.736512 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.739540 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-89k9z"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.746931 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.784978 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85d485cb5-xkqdq"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.786601 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.816883 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-6w25c" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.817115 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.817220 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.817461 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-x5tpq" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.817572 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.817698 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.817804 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.819483 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.821794 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.832664 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.833367 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.837961 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838002 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh7tw\" (UniqueName: \"kubernetes.io/projected/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-kube-api-access-lh7tw\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838026 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-scripts\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838077 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d32a26f-bd14-41a7-872b-e102ea95f696-logs\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838114 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lppbg\" (UniqueName: \"kubernetes.io/projected/9d32a26f-bd14-41a7-872b-e102ea95f696-kube-api-access-lppbg\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838135 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-config-data\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838170 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-config-data\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838196 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-combined-ca-bundle\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838213 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-logs\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838231 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838251 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838278 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fec39559-ddae-4e19-8941-0fc3209084cd-etc-machine-id\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838303 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chvrh\" (UniqueName: \"kubernetes.io/projected/fec39559-ddae-4e19-8941-0fc3209084cd-kube-api-access-chvrh\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838324 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838349 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-config-data\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.838377 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-db-sync-config-data\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.839496 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d32a26f-bd14-41a7-872b-e102ea95f696-logs\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.841309 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-logs\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.861433 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.863738 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.863821 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.878692 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-config-data\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.886350 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.902882 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-config-data\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.905246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lppbg\" (UniqueName: \"kubernetes.io/projected/9d32a26f-bd14-41a7-872b-e102ea95f696-kube-api-access-lppbg\") pod \"watcher-decision-engine-0\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.935352 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh7tw\" (UniqueName: \"kubernetes.io/projected/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-kube-api-access-lh7tw\") pod \"watcher-api-0\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " pod="openstack/watcher-api-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.938752 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.940479 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.940793 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-scripts\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.941836 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-config-data\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942156 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnpx8\" (UniqueName: \"kubernetes.io/projected/233d11ed-1edb-4e2d-b7b1-822685910821-kube-api-access-gnpx8\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942251 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-config-data\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942438 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64gbr\" (UniqueName: \"kubernetes.io/projected/4d6c59a9-358a-4f97-8c95-f1a23f3854be-kube-api-access-64gbr\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942511 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-scripts\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942603 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-combined-ca-bundle\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942685 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-config-data\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942817 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-scripts\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.942917 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fec39559-ddae-4e19-8941-0fc3209084cd-etc-machine-id\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.943018 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chvrh\" (UniqueName: \"kubernetes.io/projected/fec39559-ddae-4e19-8941-0fc3209084cd-kube-api-access-chvrh\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.943092 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.943167 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233d11ed-1edb-4e2d-b7b1-822685910821-horizon-secret-key\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.943283 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-log-httpd\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.943415 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-db-sync-config-data\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.943517 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233d11ed-1edb-4e2d-b7b1-822685910821-logs\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.944528 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-run-httpd\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.944949 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fec39559-ddae-4e19-8941-0fc3209084cd-etc-machine-id\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.945517 4750 generic.go:334] "Generic (PLEG): container finished" podID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerID="415d643fb68a9db0f410487650f656c384fd17e3e3e19205e7db418562a7617a" exitCode=0 Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.945620 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" event={"ID":"5772d728-1aa9-4b36-925a-d6d7694de34c","Type":"ContainerDied","Data":"415d643fb68a9db0f410487650f656c384fd17e3e3e19205e7db418562a7617a"} Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.958047 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.965485 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-combined-ca-bundle\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.970918 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-scripts\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.979440 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-89k9z"] Mar 09 18:47:03 crc kubenswrapper[4750]: I0309 18:47:03.995776 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.007801 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-db-sync-config-data\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.037737 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d485cb5-xkqdq"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049089 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnpx8\" (UniqueName: \"kubernetes.io/projected/233d11ed-1edb-4e2d-b7b1-822685910821-kube-api-access-gnpx8\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049147 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049191 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64gbr\" (UniqueName: \"kubernetes.io/projected/4d6c59a9-358a-4f97-8c95-f1a23f3854be-kube-api-access-64gbr\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049209 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-scripts\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049236 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-config-data\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049271 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-scripts\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049291 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049304 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k6kr\" (UniqueName: \"kubernetes.io/projected/53dae358-d7dd-4823-8369-d548e214333e-kube-api-access-2k6kr\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049338 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049355 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233d11ed-1edb-4e2d-b7b1-822685910821-horizon-secret-key\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049375 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-log-httpd\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049405 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233d11ed-1edb-4e2d-b7b1-822685910821-logs\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049422 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-run-httpd\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.049478 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-config-data\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.052062 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-config-data\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.055464 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chvrh\" (UniqueName: \"kubernetes.io/projected/fec39559-ddae-4e19-8941-0fc3209084cd-kube-api-access-chvrh\") pod \"cinder-db-sync-89k9z\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.056130 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-scripts\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.060310 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-run-httpd\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.060815 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.061441 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-log-httpd\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.066078 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-config-data\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.067999 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233d11ed-1edb-4e2d-b7b1-822685910821-logs\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.068982 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-scripts\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.071758 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-config-data\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.071859 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53dae358-d7dd-4823-8369-d548e214333e-logs\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.100361 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-89k9z" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.125882 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.143773 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.148098 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233d11ed-1edb-4e2d-b7b1-822685910821-horizon-secret-key\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.165941 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.172933 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64gbr\" (UniqueName: \"kubernetes.io/projected/4d6c59a9-358a-4f97-8c95-f1a23f3854be-kube-api-access-64gbr\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.173362 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.174293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.174323 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k6kr\" (UniqueName: \"kubernetes.io/projected/53dae358-d7dd-4823-8369-d548e214333e-kube-api-access-2k6kr\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.174444 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-config-data\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.174470 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53dae358-d7dd-4823-8369-d548e214333e-logs\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.174737 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-config-data\") pod \"ceilometer-0\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.175045 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53dae358-d7dd-4823-8369-d548e214333e-logs\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.180603 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnpx8\" (UniqueName: \"kubernetes.io/projected/233d11ed-1edb-4e2d-b7b1-822685910821-kube-api-access-gnpx8\") pod \"horizon-85d485cb5-xkqdq\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.203150 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.238405 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.240448 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-config-data\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.241236 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k6kr\" (UniqueName: \"kubernetes.io/projected/53dae358-d7dd-4823-8369-d548e214333e-kube-api-access-2k6kr\") pod \"watcher-applier-0\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.264799 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-x6m2l"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.266843 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.270694 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.272785 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.273084 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4p74c" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.333614 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-x6m2l"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.380828 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4vgf\" (UniqueName: \"kubernetes.io/projected/620fb043-252f-4bdc-8120-68a3978e1c82-kube-api-access-n4vgf\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.380942 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-combined-ca-bundle\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.380983 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-config\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.440814 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d55fc75b5-pjhlj"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.443259 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.452387 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.484837 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-combined-ca-bundle\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.494014 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-config\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.494600 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4vgf\" (UniqueName: \"kubernetes.io/projected/620fb043-252f-4bdc-8120-68a3978e1c82-kube-api-access-n4vgf\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.487874 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d55fc75b5-pjhlj"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.569244 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-combined-ca-bundle\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.569347 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-config\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.569358 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.601053 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4vgf\" (UniqueName: \"kubernetes.io/projected/620fb043-252f-4bdc-8120-68a3978e1c82-kube-api-access-n4vgf\") pod \"neutron-db-sync-x6m2l\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.616930 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89cbc8c6f-jpdk9"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.633267 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.673805 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-59wvm"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.677412 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-config-data\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.677502 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-scripts\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.677572 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngq5b\" (UniqueName: \"kubernetes.io/projected/bdb93bcb-fc2c-4480-85a4-709abaf41138-kube-api-access-ngq5b\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.679054 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.726234 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdb93bcb-fc2c-4480-85a4-709abaf41138-horizon-secret-key\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.726289 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb93bcb-fc2c-4480-85a4-709abaf41138-logs\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.736686 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pq62m" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.736904 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.759814 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hsgm7"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.765363 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.768427 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.768710 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xzm7g" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.768862 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.828486 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdb93bcb-fc2c-4480-85a4-709abaf41138-horizon-secret-key\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.828544 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb93bcb-fc2c-4480-85a4-709abaf41138-logs\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.828587 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbxrg\" (UniqueName: \"kubernetes.io/projected/215dfd5c-6d06-42ef-89c7-6d40910fbf28-kube-api-access-hbxrg\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.828620 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-combined-ca-bundle\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.828667 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-config-data\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.828691 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-scripts\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.828705 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngq5b\" (UniqueName: \"kubernetes.io/projected/bdb93bcb-fc2c-4480-85a4-709abaf41138-kube-api-access-ngq5b\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.829220 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-db-sync-config-data\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.830401 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb93bcb-fc2c-4480-85a4-709abaf41138-logs\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.830995 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-scripts\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.831817 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-config-data\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.835659 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdb93bcb-fc2c-4480-85a4-709abaf41138-horizon-secret-key\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.841824 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.843889 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.847127 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qbg7p" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.847861 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.848196 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.849337 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f5cd9697c-zjwc7"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.849787 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.850948 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.857108 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngq5b\" (UniqueName: \"kubernetes.io/projected/bdb93bcb-fc2c-4480-85a4-709abaf41138-kube-api-access-ngq5b\") pod \"horizon-d55fc75b5-pjhlj\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.871918 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-59wvm"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.889610 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.899207 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hsgm7"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.914283 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f5cd9697c-zjwc7"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933506 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-combined-ca-bundle\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933559 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-db-sync-config-data\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933599 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2afe8078-9724-4ab3-b5ee-79c978ce9bea-logs\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933615 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz94p\" (UniqueName: \"kubernetes.io/projected/2afe8078-9724-4ab3-b5ee-79c978ce9bea-kube-api-access-rz94p\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933678 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbxrg\" (UniqueName: \"kubernetes.io/projected/215dfd5c-6d06-42ef-89c7-6d40910fbf28-kube-api-access-hbxrg\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933707 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-combined-ca-bundle\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933763 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-scripts\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.933792 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-config-data\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.934943 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.936866 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-db-sync-config-data\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.942328 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-combined-ca-bundle\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.948586 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.950852 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.954834 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.955282 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.955438 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbxrg\" (UniqueName: \"kubernetes.io/projected/215dfd5c-6d06-42ef-89c7-6d40910fbf28-kube-api-access-hbxrg\") pod \"barbican-db-sync-59wvm\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:04 crc kubenswrapper[4750]: I0309 18:47:04.958041 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035712 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnnbn\" (UniqueName: \"kubernetes.io/projected/031456d8-c16a-49fa-9b19-b22060292ac3-kube-api-access-rnnbn\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035765 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-svc\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035792 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4nvl\" (UniqueName: \"kubernetes.io/projected/189e2dcb-e49f-4af8-8907-807a81f5dfd3-kube-api-access-p4nvl\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035812 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035834 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035849 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035870 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-scripts\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035893 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-config\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035915 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035934 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-config-data\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035956 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035974 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.035990 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-combined-ca-bundle\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.036010 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-logs\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.036024 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.036058 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2afe8078-9724-4ab3-b5ee-79c978ce9bea-logs\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.036073 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz94p\" (UniqueName: \"kubernetes.io/projected/2afe8078-9724-4ab3-b5ee-79c978ce9bea-kube-api-access-rz94p\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.036109 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.036136 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.037834 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2afe8078-9724-4ab3-b5ee-79c978ce9bea-logs\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.044228 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89cbc8c6f-jpdk9"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.047140 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-scripts\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.047221 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-config-data\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.047373 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-combined-ca-bundle\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.052140 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6kmdp"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.061699 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz94p\" (UniqueName: \"kubernetes.io/projected/2afe8078-9724-4ab3-b5ee-79c978ce9bea-kube-api-access-rz94p\") pod \"placement-db-sync-hsgm7\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.105198 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.120798 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.138574 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.138896 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnnbn\" (UniqueName: \"kubernetes.io/projected/031456d8-c16a-49fa-9b19-b22060292ac3-kube-api-access-rnnbn\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.138943 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-svc\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.138974 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4nvl\" (UniqueName: \"kubernetes.io/projected/189e2dcb-e49f-4af8-8907-807a81f5dfd3-kube-api-access-p4nvl\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139008 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139041 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139063 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct6x4\" (UniqueName: \"kubernetes.io/projected/8573c870-7c80-4500-8e66-3ee2a6088b66-kube-api-access-ct6x4\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139081 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139115 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-config\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139145 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139198 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139217 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139243 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-logs\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139260 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139282 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-scripts\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139307 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-logs\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139329 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139388 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-config-data\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139420 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139440 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139504 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.139826 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.140650 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-swift-storage-0\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.140866 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-svc\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.141268 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.141843 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.141925 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.142071 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-logs\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.142652 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-config\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.147169 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.150539 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.156376 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.159048 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.166482 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4nvl\" (UniqueName: \"kubernetes.io/projected/189e2dcb-e49f-4af8-8907-807a81f5dfd3-kube-api-access-p4nvl\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.166549 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnnbn\" (UniqueName: \"kubernetes.io/projected/031456d8-c16a-49fa-9b19-b22060292ac3-kube-api-access-rnnbn\") pod \"dnsmasq-dns-5f5cd9697c-zjwc7\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.183192 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.200076 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.245906 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-scripts\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.245956 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-logs\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.245984 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.246054 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-config-data\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.246082 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.246119 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.246215 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct6x4\" (UniqueName: \"kubernetes.io/projected/8573c870-7c80-4500-8e66-3ee2a6088b66-kube-api-access-ct6x4\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.246271 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.249352 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.249554 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-logs\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.251931 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.253522 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.264039 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-config-data\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.272452 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-scripts\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.273207 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.282522 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct6x4\" (UniqueName: \"kubernetes.io/projected/8573c870-7c80-4500-8e66-3ee2a6088b66-kube-api-access-ct6x4\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.351120 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.479378 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.540487 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-89k9z"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.582359 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.590719 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.641808 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.686230 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.837948 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.863061 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d485cb5-xkqdq"] Mar 09 18:47:05 crc kubenswrapper[4750]: W0309 18:47:05.874881 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod233d11ed_1edb_4e2d_b7b1_822685910821.slice/crio-024a6808b63b4f55752324613cf87d6011eb796cd6dd0b519854853fd6216488 WatchSource:0}: Error finding container 024a6808b63b4f55752324613cf87d6011eb796cd6dd0b519854853fd6216488: Status 404 returned error can't find the container with id 024a6808b63b4f55752324613cf87d6011eb796cd6dd0b519854853fd6216488 Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.882200 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.974114 4750 generic.go:334] "Generic (PLEG): container finished" podID="3f1158dc-0354-469d-82f7-1d5e1a9de48f" containerID="19a05e52980dfb24ede00db72ae87dafe40e2a8cb62d6ad2aa5d5b2c164cf6a6" exitCode=0 Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.974192 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" event={"ID":"3f1158dc-0354-469d-82f7-1d5e1a9de48f","Type":"ContainerDied","Data":"19a05e52980dfb24ede00db72ae87dafe40e2a8cb62d6ad2aa5d5b2c164cf6a6"} Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.974225 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" event={"ID":"3f1158dc-0354-469d-82f7-1d5e1a9de48f","Type":"ContainerStarted","Data":"d2947de05a2804f29b613e5f5f79a1ef5cac388fa113fca6db62d2aeede3fbe0"} Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.988933 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6kmdp" event={"ID":"60a57822-e425-4ef7-8da8-f83a0c0abe60","Type":"ContainerStarted","Data":"209d4a9e62894dda18feac538f1f3c0dd3d0ff35260741bfb279223d7d96bcbd"} Mar 09 18:47:05 crc kubenswrapper[4750]: I0309 18:47:05.989022 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6kmdp" event={"ID":"60a57822-e425-4ef7-8da8-f83a0c0abe60","Type":"ContainerStarted","Data":"e38d1e2950cd59387d2922bc962833f16daf4b900241a0a5f631c3a94b929e16"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.004806 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-89k9z" event={"ID":"fec39559-ddae-4e19-8941-0fc3209084cd","Type":"ContainerStarted","Data":"5fdc9f515d14510136d8dd08d3bc3d6b17f383963220bd57556a0e715ee72f41"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.012773 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" event={"ID":"5772d728-1aa9-4b36-925a-d6d7694de34c","Type":"ContainerDied","Data":"2d8be5663a529e390f7c12be785803d7e253d5195777096ab8a780da17ee00da"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.012841 4750 scope.go:117] "RemoveContainer" containerID="415d643fb68a9db0f410487650f656c384fd17e3e3e19205e7db418562a7617a" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.013011 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f8c47bdf-hmx4v" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.022368 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6kmdp" podStartSLOduration=3.022348577 podStartE2EDuration="3.022348577s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:06.018093532 +0000 UTC m=+1307.360565930" watchObservedRunningTime="2026-03-09 18:47:06.022348577 +0000 UTC m=+1307.364820975" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.036670 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerStarted","Data":"fc29b5100f86212c229e7035c2d0bc971b1f509ff517b56860b5e2f0a1900351"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.039232 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"53dae358-d7dd-4823-8369-d548e214333e","Type":"ContainerStarted","Data":"a9b4eddf4191e6e62d4b90b10776648af8ca32f0e90a7242f84ef0704cf64949"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.041302 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"9d32a26f-bd14-41a7-872b-e102ea95f696","Type":"ContainerStarted","Data":"448fbabe0dcd93888c64acd03db8f6a301f57c5a438fb54c8e58ea16bd2b3336"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.046959 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d485cb5-xkqdq" event={"ID":"233d11ed-1edb-4e2d-b7b1-822685910821","Type":"ContainerStarted","Data":"024a6808b63b4f55752324613cf87d6011eb796cd6dd0b519854853fd6216488"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.058391 4750 scope.go:117] "RemoveContainer" containerID="154ff451a12cb04e1f3e6673e5f88d9b1989498db68b23f07de84514fbf4a063" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.072479 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-sb\") pod \"5772d728-1aa9-4b36-925a-d6d7694de34c\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.072600 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-config\") pod \"5772d728-1aa9-4b36-925a-d6d7694de34c\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.072679 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-svc\") pod \"5772d728-1aa9-4b36-925a-d6d7694de34c\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.072720 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9gk9\" (UniqueName: \"kubernetes.io/projected/5772d728-1aa9-4b36-925a-d6d7694de34c-kube-api-access-t9gk9\") pod \"5772d728-1aa9-4b36-925a-d6d7694de34c\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.072833 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-swift-storage-0\") pod \"5772d728-1aa9-4b36-925a-d6d7694de34c\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.072879 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-nb\") pod \"5772d728-1aa9-4b36-925a-d6d7694de34c\" (UID: \"5772d728-1aa9-4b36-925a-d6d7694de34c\") " Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.078348 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433","Type":"ContainerStarted","Data":"ba7153ca4c17724d4bdfc1dbb6a92741a5124afee90649f64118c94683a4c84c"} Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.096456 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5772d728-1aa9-4b36-925a-d6d7694de34c-kube-api-access-t9gk9" (OuterVolumeSpecName: "kube-api-access-t9gk9") pod "5772d728-1aa9-4b36-925a-d6d7694de34c" (UID: "5772d728-1aa9-4b36-925a-d6d7694de34c"). InnerVolumeSpecName "kube-api-access-t9gk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.139078 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5772d728-1aa9-4b36-925a-d6d7694de34c" (UID: "5772d728-1aa9-4b36-925a-d6d7694de34c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.180291 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9gk9\" (UniqueName: \"kubernetes.io/projected/5772d728-1aa9-4b36-925a-d6d7694de34c-kube-api-access-t9gk9\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.182105 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.185577 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5772d728-1aa9-4b36-925a-d6d7694de34c" (UID: "5772d728-1aa9-4b36-925a-d6d7694de34c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.200544 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5772d728-1aa9-4b36-925a-d6d7694de34c" (UID: "5772d728-1aa9-4b36-925a-d6d7694de34c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.203186 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d55fc75b5-pjhlj"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.206126 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-config" (OuterVolumeSpecName: "config") pod "5772d728-1aa9-4b36-925a-d6d7694de34c" (UID: "5772d728-1aa9-4b36-925a-d6d7694de34c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.231537 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-x6m2l"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.252976 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.284597 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.284636 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.284645 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.304715 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5772d728-1aa9-4b36-925a-d6d7694de34c" (UID: "5772d728-1aa9-4b36-925a-d6d7694de34c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.373484 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.384361 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85d485cb5-xkqdq"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.386813 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5772d728-1aa9-4b36-925a-d6d7694de34c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.418737 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f8c47bdf-hmx4v"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.439019 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85f8c47bdf-hmx4v"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.457000 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hsgm7"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.486428 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-59wvm"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.502117 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f5cd9697c-zjwc7"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.517215 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6dcf6cd949-lbxvv"] Mar 09 18:47:06 crc kubenswrapper[4750]: E0309 18:47:06.517650 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerName="dnsmasq-dns" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.517666 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerName="dnsmasq-dns" Mar 09 18:47:06 crc kubenswrapper[4750]: E0309 18:47:06.517688 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerName="init" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.517696 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerName="init" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.518197 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5772d728-1aa9-4b36-925a-d6d7694de34c" containerName="dnsmasq-dns" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.520552 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.527665 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.544398 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6dcf6cd949-lbxvv"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.692143 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-config-data\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.692475 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55x5z\" (UniqueName: \"kubernetes.io/projected/66e0e816-a509-43e6-824e-960e537d6bfc-kube-api-access-55x5z\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.692516 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66e0e816-a509-43e6-824e-960e537d6bfc-logs\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.692569 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66e0e816-a509-43e6-824e-960e537d6bfc-horizon-secret-key\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.693719 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-scripts\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.795079 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.795263 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55x5z\" (UniqueName: \"kubernetes.io/projected/66e0e816-a509-43e6-824e-960e537d6bfc-kube-api-access-55x5z\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.795312 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66e0e816-a509-43e6-824e-960e537d6bfc-logs\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.795364 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66e0e816-a509-43e6-824e-960e537d6bfc-horizon-secret-key\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.795459 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-scripts\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.795505 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-config-data\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.798543 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66e0e816-a509-43e6-824e-960e537d6bfc-logs\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.799675 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-scripts\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.802422 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-config-data\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.829423 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55x5z\" (UniqueName: \"kubernetes.io/projected/66e0e816-a509-43e6-824e-960e537d6bfc-kube-api-access-55x5z\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.876592 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.823509 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66e0e816-a509-43e6-824e-960e537d6bfc-horizon-secret-key\") pod \"horizon-6dcf6cd949-lbxvv\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.916081 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.924294 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.999641 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-config\") pod \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " Mar 09 18:47:06 crc kubenswrapper[4750]: I0309 18:47:06.999774 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-swift-storage-0\") pod \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:06.999856 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-nb\") pod \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:06.999891 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nm5k\" (UniqueName: \"kubernetes.io/projected/3f1158dc-0354-469d-82f7-1d5e1a9de48f-kube-api-access-9nm5k\") pod \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:06.999956 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-sb\") pod \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.000000 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-svc\") pod \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\" (UID: \"3f1158dc-0354-469d-82f7-1d5e1a9de48f\") " Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.006529 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.027615 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f1158dc-0354-469d-82f7-1d5e1a9de48f-kube-api-access-9nm5k" (OuterVolumeSpecName: "kube-api-access-9nm5k") pod "3f1158dc-0354-469d-82f7-1d5e1a9de48f" (UID: "3f1158dc-0354-469d-82f7-1d5e1a9de48f"). InnerVolumeSpecName "kube-api-access-9nm5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.288163 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nm5k\" (UniqueName: \"kubernetes.io/projected/3f1158dc-0354-469d-82f7-1d5e1a9de48f-kube-api-access-9nm5k\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.294481 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-config" (OuterVolumeSpecName: "config") pod "3f1158dc-0354-469d-82f7-1d5e1a9de48f" (UID: "3f1158dc-0354-469d-82f7-1d5e1a9de48f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.322251 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189e2dcb-e49f-4af8-8907-807a81f5dfd3","Type":"ContainerStarted","Data":"317bd1028df012ad0e9762fdc70b7b94a7cf5239ca6c6ff25dc17b201ac4f7f8"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.323750 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8573c870-7c80-4500-8e66-3ee2a6088b66","Type":"ContainerStarted","Data":"1288f158988b1e47e72a991e7fb035688a93eef51e57e58d7a327f0a4248d135"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.325762 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" event={"ID":"031456d8-c16a-49fa-9b19-b22060292ac3","Type":"ContainerStarted","Data":"f531f7fe33696cfd23d1d0cf3fbef422e8892bd1ff13db52684a150209325314"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.340273 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-x6m2l" event={"ID":"620fb043-252f-4bdc-8120-68a3978e1c82","Type":"ContainerStarted","Data":"79d793d41819e39872483071d0df21fac11e13760912f8b4ee73535605ed9d8a"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.340320 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-x6m2l" event={"ID":"620fb043-252f-4bdc-8120-68a3978e1c82","Type":"ContainerStarted","Data":"0d71fc431522ae1417fec7191123adf4c3004de0bde72099946fcc8c45aecb27"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.343312 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" event={"ID":"3f1158dc-0354-469d-82f7-1d5e1a9de48f","Type":"ContainerDied","Data":"d2947de05a2804f29b613e5f5f79a1ef5cac388fa113fca6db62d2aeede3fbe0"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.343362 4750 scope.go:117] "RemoveContainer" containerID="19a05e52980dfb24ede00db72ae87dafe40e2a8cb62d6ad2aa5d5b2c164cf6a6" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.343334 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89cbc8c6f-jpdk9" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.345511 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433","Type":"ContainerStarted","Data":"772d7a0f90ac442db02b04642c9d04d45eef711be2787a15008d14446a3959ee"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.345551 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433","Type":"ContainerStarted","Data":"40bdcea7bb85288f1c4bbb741b609100cb12caba25eee36bab891833ab9fec41"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.345694 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api-log" containerID="cri-o://40bdcea7bb85288f1c4bbb741b609100cb12caba25eee36bab891833ab9fec41" gracePeriod=30 Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.346219 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.346230 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api" containerID="cri-o://772d7a0f90ac442db02b04642c9d04d45eef711be2787a15008d14446a3959ee" gracePeriod=30 Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.350512 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d55fc75b5-pjhlj" event={"ID":"bdb93bcb-fc2c-4480-85a4-709abaf41138","Type":"ContainerStarted","Data":"404bc2c3426daa156c2a7df3e8d29207f6b67aef7935fe69cbe4e298e98c99c7"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.352009 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsgm7" event={"ID":"2afe8078-9724-4ab3-b5ee-79c978ce9bea","Type":"ContainerStarted","Data":"d9ddcbfb1c220ea4f30459640ea4b9f2c20d0557c721a088be9aa53cbc747177"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.354103 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-59wvm" event={"ID":"215dfd5c-6d06-42ef-89c7-6d40910fbf28","Type":"ContainerStarted","Data":"5308a9c506d492aaa4a052237cfde65a23247123a3f9d9159f5aa133183ca01b"} Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.390453 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.397492 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5772d728-1aa9-4b36-925a-d6d7694de34c" path="/var/lib/kubelet/pods/5772d728-1aa9-4b36-925a-d6d7694de34c/volumes" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.615444 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=4.615422068 podStartE2EDuration="4.615422068s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:07.612993372 +0000 UTC m=+1308.955465770" watchObservedRunningTime="2026-03-09 18:47:07.615422068 +0000 UTC m=+1308.957894466" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.619436 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-x6m2l" podStartSLOduration=4.619421847 podStartE2EDuration="4.619421847s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:07.566747146 +0000 UTC m=+1308.909219584" watchObservedRunningTime="2026-03-09 18:47:07.619421847 +0000 UTC m=+1308.961894245" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.676259 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3f1158dc-0354-469d-82f7-1d5e1a9de48f" (UID: "3f1158dc-0354-469d-82f7-1d5e1a9de48f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.703387 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.715010 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3f1158dc-0354-469d-82f7-1d5e1a9de48f" (UID: "3f1158dc-0354-469d-82f7-1d5e1a9de48f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.778798 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3f1158dc-0354-469d-82f7-1d5e1a9de48f" (UID: "3f1158dc-0354-469d-82f7-1d5e1a9de48f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.781543 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3f1158dc-0354-469d-82f7-1d5e1a9de48f" (UID: "3f1158dc-0354-469d-82f7-1d5e1a9de48f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.806750 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.807007 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.807087 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f1158dc-0354-469d-82f7-1d5e1a9de48f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:07 crc kubenswrapper[4750]: I0309 18:47:07.856951 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6dcf6cd949-lbxvv"] Mar 09 18:47:07 crc kubenswrapper[4750]: W0309 18:47:07.905620 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66e0e816_a509_43e6_824e_960e537d6bfc.slice/crio-3c57aa10969e8c7a8a2febcb1460bbaa9e3895235aa005486f7577e2222b0075 WatchSource:0}: Error finding container 3c57aa10969e8c7a8a2febcb1460bbaa9e3895235aa005486f7577e2222b0075: Status 404 returned error can't find the container with id 3c57aa10969e8c7a8a2febcb1460bbaa9e3895235aa005486f7577e2222b0075 Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.016072 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89cbc8c6f-jpdk9"] Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.034135 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89cbc8c6f-jpdk9"] Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.417072 4750 generic.go:334] "Generic (PLEG): container finished" podID="031456d8-c16a-49fa-9b19-b22060292ac3" containerID="53113e128ce42ad1d1f8b5c4174cd3215f52c58caa7e388816634f12bc81f2db" exitCode=0 Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.417181 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" event={"ID":"031456d8-c16a-49fa-9b19-b22060292ac3","Type":"ContainerDied","Data":"53113e128ce42ad1d1f8b5c4174cd3215f52c58caa7e388816634f12bc81f2db"} Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.444906 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189e2dcb-e49f-4af8-8907-807a81f5dfd3","Type":"ContainerStarted","Data":"42c659a821382be1ef3d4981241e21783ff08f6a4ea3b8ac936124dafe96e0ac"} Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.467024 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dcf6cd949-lbxvv" event={"ID":"66e0e816-a509-43e6-824e-960e537d6bfc","Type":"ContainerStarted","Data":"3c57aa10969e8c7a8a2febcb1460bbaa9e3895235aa005486f7577e2222b0075"} Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.471887 4750 generic.go:334] "Generic (PLEG): container finished" podID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerID="40bdcea7bb85288f1c4bbb741b609100cb12caba25eee36bab891833ab9fec41" exitCode=143 Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.472786 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433","Type":"ContainerDied","Data":"40bdcea7bb85288f1c4bbb741b609100cb12caba25eee36bab891833ab9fec41"} Mar 09 18:47:08 crc kubenswrapper[4750]: I0309 18:47:08.997802 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:47:09 crc kubenswrapper[4750]: I0309 18:47:09.450379 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f1158dc-0354-469d-82f7-1d5e1a9de48f" path="/var/lib/kubelet/pods/3f1158dc-0354-469d-82f7-1d5e1a9de48f/volumes" Mar 09 18:47:09 crc kubenswrapper[4750]: I0309 18:47:09.544976 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8573c870-7c80-4500-8e66-3ee2a6088b66","Type":"ContainerStarted","Data":"bf0754d343051a78eff0c4d13faffa6e7f84a2ad8cae7fc86bae0da0abc4156a"} Mar 09 18:47:09 crc kubenswrapper[4750]: I0309 18:47:09.571367 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" event={"ID":"031456d8-c16a-49fa-9b19-b22060292ac3","Type":"ContainerStarted","Data":"4532b42f7be8a187289ed33dac72c539a14b331df2223ee86ccbbd600a95d28d"} Mar 09 18:47:09 crc kubenswrapper[4750]: I0309 18:47:09.571584 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:09 crc kubenswrapper[4750]: I0309 18:47:09.741646 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" podStartSLOduration=5.741609564 podStartE2EDuration="5.741609564s" podCreationTimestamp="2026-03-09 18:47:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:09.738055437 +0000 UTC m=+1311.080527835" watchObservedRunningTime="2026-03-09 18:47:09.741609564 +0000 UTC m=+1311.084081952" Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.591826 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189e2dcb-e49f-4af8-8907-807a81f5dfd3","Type":"ContainerStarted","Data":"5ea9b1043fc3d3c485d640964e713fd62560bdcf4e3e7584547370dd29b722de"} Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.592196 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-log" containerID="cri-o://42c659a821382be1ef3d4981241e21783ff08f6a4ea3b8ac936124dafe96e0ac" gracePeriod=30 Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.592342 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-httpd" containerID="cri-o://5ea9b1043fc3d3c485d640964e713fd62560bdcf4e3e7584547370dd29b722de" gracePeriod=30 Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.595854 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-log" containerID="cri-o://bf0754d343051a78eff0c4d13faffa6e7f84a2ad8cae7fc86bae0da0abc4156a" gracePeriod=30 Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.595990 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-httpd" containerID="cri-o://e721b26c39b4bd7db5346d1e67ec0fade4b05f0b0caaad98c67c833c88f5ca4f" gracePeriod=30 Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.596075 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8573c870-7c80-4500-8e66-3ee2a6088b66","Type":"ContainerStarted","Data":"e721b26c39b4bd7db5346d1e67ec0fade4b05f0b0caaad98c67c833c88f5ca4f"} Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.626192 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.626167615 podStartE2EDuration="6.626167615s" podCreationTimestamp="2026-03-09 18:47:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:10.613048959 +0000 UTC m=+1311.955521377" watchObservedRunningTime="2026-03-09 18:47:10.626167615 +0000 UTC m=+1311.968640013" Mar 09 18:47:10 crc kubenswrapper[4750]: I0309 18:47:10.642755 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.642736445 podStartE2EDuration="6.642736445s" podCreationTimestamp="2026-03-09 18:47:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:10.641201533 +0000 UTC m=+1311.983673941" watchObservedRunningTime="2026-03-09 18:47:10.642736445 +0000 UTC m=+1311.985208843" Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.550964 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.607100 4750 generic.go:334] "Generic (PLEG): container finished" podID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerID="5ea9b1043fc3d3c485d640964e713fd62560bdcf4e3e7584547370dd29b722de" exitCode=0 Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.607150 4750 generic.go:334] "Generic (PLEG): container finished" podID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerID="42c659a821382be1ef3d4981241e21783ff08f6a4ea3b8ac936124dafe96e0ac" exitCode=143 Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.607188 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189e2dcb-e49f-4af8-8907-807a81f5dfd3","Type":"ContainerDied","Data":"5ea9b1043fc3d3c485d640964e713fd62560bdcf4e3e7584547370dd29b722de"} Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.607252 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189e2dcb-e49f-4af8-8907-807a81f5dfd3","Type":"ContainerDied","Data":"42c659a821382be1ef3d4981241e21783ff08f6a4ea3b8ac936124dafe96e0ac"} Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.610479 4750 generic.go:334] "Generic (PLEG): container finished" podID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerID="e721b26c39b4bd7db5346d1e67ec0fade4b05f0b0caaad98c67c833c88f5ca4f" exitCode=0 Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.610499 4750 generic.go:334] "Generic (PLEG): container finished" podID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerID="bf0754d343051a78eff0c4d13faffa6e7f84a2ad8cae7fc86bae0da0abc4156a" exitCode=143 Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.610523 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8573c870-7c80-4500-8e66-3ee2a6088b66","Type":"ContainerDied","Data":"e721b26c39b4bd7db5346d1e67ec0fade4b05f0b0caaad98c67c833c88f5ca4f"} Mar 09 18:47:11 crc kubenswrapper[4750]: I0309 18:47:11.610549 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8573c870-7c80-4500-8e66-3ee2a6088b66","Type":"ContainerDied","Data":"bf0754d343051a78eff0c4d13faffa6e7f84a2ad8cae7fc86bae0da0abc4156a"} Mar 09 18:47:12 crc kubenswrapper[4750]: I0309 18:47:12.623674 4750 generic.go:334] "Generic (PLEG): container finished" podID="60a57822-e425-4ef7-8da8-f83a0c0abe60" containerID="209d4a9e62894dda18feac538f1f3c0dd3d0ff35260741bfb279223d7d96bcbd" exitCode=0 Mar 09 18:47:12 crc kubenswrapper[4750]: I0309 18:47:12.623765 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6kmdp" event={"ID":"60a57822-e425-4ef7-8da8-f83a0c0abe60","Type":"ContainerDied","Data":"209d4a9e62894dda18feac538f1f3c0dd3d0ff35260741bfb279223d7d96bcbd"} Mar 09 18:47:12 crc kubenswrapper[4750]: I0309 18:47:12.940966 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d55fc75b5-pjhlj"] Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.006180 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5ddc4b4b8d-f5q7t"] Mar 09 18:47:13 crc kubenswrapper[4750]: E0309 18:47:13.006721 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1158dc-0354-469d-82f7-1d5e1a9de48f" containerName="init" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.006736 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1158dc-0354-469d-82f7-1d5e1a9de48f" containerName="init" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.006918 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f1158dc-0354-469d-82f7-1d5e1a9de48f" containerName="init" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.008358 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.012655 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.033481 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ddc4b4b8d-f5q7t"] Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.065959 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6dcf6cd949-lbxvv"] Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.102300 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fdd5dd964-x7n8q"] Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.104401 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.122808 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fdd5dd964-x7n8q"] Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.131246 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qs7d\" (UniqueName: \"kubernetes.io/projected/ef45ec45-1ef2-46e6-bf33-0c092885679f-kube-api-access-6qs7d\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.131304 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-secret-key\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.131368 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-combined-ca-bundle\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.131413 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-config-data\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.131485 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef45ec45-1ef2-46e6-bf33-0c092885679f-logs\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.131507 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-tls-certs\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.131589 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-scripts\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235228 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-config-data\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235323 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0509f0f1-226d-4626-998d-b9065a8a634b-config-data\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235426 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0509f0f1-226d-4626-998d-b9065a8a634b-logs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235469 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef45ec45-1ef2-46e6-bf33-0c092885679f-logs\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235493 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-tls-certs\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235535 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-horizon-tls-certs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235560 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-combined-ca-bundle\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235596 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-horizon-secret-key\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235721 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-scripts\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235761 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ssfs\" (UniqueName: \"kubernetes.io/projected/0509f0f1-226d-4626-998d-b9065a8a634b-kube-api-access-8ssfs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235794 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qs7d\" (UniqueName: \"kubernetes.io/projected/ef45ec45-1ef2-46e6-bf33-0c092885679f-kube-api-access-6qs7d\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235866 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-secret-key\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235946 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-combined-ca-bundle\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.235975 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0509f0f1-226d-4626-998d-b9065a8a634b-scripts\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.236389 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef45ec45-1ef2-46e6-bf33-0c092885679f-logs\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.236536 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-config-data\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.237090 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-scripts\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.245279 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-tls-certs\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.245463 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-secret-key\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.245533 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-combined-ca-bundle\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.253557 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qs7d\" (UniqueName: \"kubernetes.io/projected/ef45ec45-1ef2-46e6-bf33-0c092885679f-kube-api-access-6qs7d\") pod \"horizon-5ddc4b4b8d-f5q7t\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.337676 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ssfs\" (UniqueName: \"kubernetes.io/projected/0509f0f1-226d-4626-998d-b9065a8a634b-kube-api-access-8ssfs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.337771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0509f0f1-226d-4626-998d-b9065a8a634b-scripts\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.337825 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0509f0f1-226d-4626-998d-b9065a8a634b-config-data\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.337856 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0509f0f1-226d-4626-998d-b9065a8a634b-logs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.337893 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-horizon-tls-certs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.337911 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-combined-ca-bundle\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.337937 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-horizon-secret-key\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.338622 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0509f0f1-226d-4626-998d-b9065a8a634b-logs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.338948 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0509f0f1-226d-4626-998d-b9065a8a634b-scripts\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.340248 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0509f0f1-226d-4626-998d-b9065a8a634b-config-data\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.341750 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-combined-ca-bundle\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.343550 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-horizon-tls-certs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.344971 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0509f0f1-226d-4626-998d-b9065a8a634b-horizon-secret-key\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.348109 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.357410 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ssfs\" (UniqueName: \"kubernetes.io/projected/0509f0f1-226d-4626-998d-b9065a8a634b-kube-api-access-8ssfs\") pod \"horizon-7fdd5dd964-x7n8q\" (UID: \"0509f0f1-226d-4626-998d-b9065a8a634b\") " pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:13 crc kubenswrapper[4750]: I0309 18:47:13.436717 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:15 crc kubenswrapper[4750]: I0309 18:47:15.201898 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:15 crc kubenswrapper[4750]: I0309 18:47:15.270842 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cd9f6b887-2nlmb"] Mar 09 18:47:15 crc kubenswrapper[4750]: I0309 18:47:15.271354 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" containerID="cri-o://49fbad12eb84b0f3d14f848f220cc82823fd7871d7abf0b485264235d15dc73f" gracePeriod=10 Mar 09 18:47:15 crc kubenswrapper[4750]: I0309 18:47:15.675118 4750 generic.go:334] "Generic (PLEG): container finished" podID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerID="49fbad12eb84b0f3d14f848f220cc82823fd7871d7abf0b485264235d15dc73f" exitCode=0 Mar 09 18:47:15 crc kubenswrapper[4750]: I0309 18:47:15.675435 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" event={"ID":"66f23451-0dce-4bfa-95e7-8d17adc315d0","Type":"ContainerDied","Data":"49fbad12eb84b0f3d14f848f220cc82823fd7871d7abf0b485264235d15dc73f"} Mar 09 18:47:16 crc kubenswrapper[4750]: I0309 18:47:16.961951 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.013694 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.064670 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-scripts\") pod \"60a57822-e425-4ef7-8da8-f83a0c0abe60\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.064745 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-combined-ca-bundle\") pod \"60a57822-e425-4ef7-8da8-f83a0c0abe60\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.064766 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh89x\" (UniqueName: \"kubernetes.io/projected/60a57822-e425-4ef7-8da8-f83a0c0abe60-kube-api-access-sh89x\") pod \"60a57822-e425-4ef7-8da8-f83a0c0abe60\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.064798 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-config-data\") pod \"60a57822-e425-4ef7-8da8-f83a0c0abe60\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.064848 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-fernet-keys\") pod \"60a57822-e425-4ef7-8da8-f83a0c0abe60\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.064898 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-credential-keys\") pod \"60a57822-e425-4ef7-8da8-f83a0c0abe60\" (UID: \"60a57822-e425-4ef7-8da8-f83a0c0abe60\") " Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.071386 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "60a57822-e425-4ef7-8da8-f83a0c0abe60" (UID: "60a57822-e425-4ef7-8da8-f83a0c0abe60"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.072165 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "60a57822-e425-4ef7-8da8-f83a0c0abe60" (UID: "60a57822-e425-4ef7-8da8-f83a0c0abe60"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.072790 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a57822-e425-4ef7-8da8-f83a0c0abe60-kube-api-access-sh89x" (OuterVolumeSpecName: "kube-api-access-sh89x") pod "60a57822-e425-4ef7-8da8-f83a0c0abe60" (UID: "60a57822-e425-4ef7-8da8-f83a0c0abe60"). InnerVolumeSpecName "kube-api-access-sh89x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.073759 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-scripts" (OuterVolumeSpecName: "scripts") pod "60a57822-e425-4ef7-8da8-f83a0c0abe60" (UID: "60a57822-e425-4ef7-8da8-f83a0c0abe60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.103770 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-config-data" (OuterVolumeSpecName: "config-data") pod "60a57822-e425-4ef7-8da8-f83a0c0abe60" (UID: "60a57822-e425-4ef7-8da8-f83a0c0abe60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.111314 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60a57822-e425-4ef7-8da8-f83a0c0abe60" (UID: "60a57822-e425-4ef7-8da8-f83a0c0abe60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.166548 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.166605 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.166617 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh89x\" (UniqueName: \"kubernetes.io/projected/60a57822-e425-4ef7-8da8-f83a0c0abe60-kube-api-access-sh89x\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.166640 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.166649 4750 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.166660 4750 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60a57822-e425-4ef7-8da8-f83a0c0abe60-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.707147 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6kmdp" event={"ID":"60a57822-e425-4ef7-8da8-f83a0c0abe60","Type":"ContainerDied","Data":"e38d1e2950cd59387d2922bc962833f16daf4b900241a0a5f631c3a94b929e16"} Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.707411 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e38d1e2950cd59387d2922bc962833f16daf4b900241a0a5f631c3a94b929e16" Mar 09 18:47:18 crc kubenswrapper[4750]: I0309 18:47:18.707418 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6kmdp" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.112813 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6kmdp"] Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.124774 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6kmdp"] Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.207849 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t8lkt"] Mar 09 18:47:19 crc kubenswrapper[4750]: E0309 18:47:19.208226 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a57822-e425-4ef7-8da8-f83a0c0abe60" containerName="keystone-bootstrap" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.208243 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a57822-e425-4ef7-8da8-f83a0c0abe60" containerName="keystone-bootstrap" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.208756 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a57822-e425-4ef7-8da8-f83a0c0abe60" containerName="keystone-bootstrap" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.209984 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.212396 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.212652 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.213491 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.213704 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9flmk" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.215733 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.230035 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t8lkt"] Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.292338 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-fernet-keys\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.292509 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-credential-keys\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.292581 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-combined-ca-bundle\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.292727 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-scripts\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.292817 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-config-data\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.292863 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfwf6\" (UniqueName: \"kubernetes.io/projected/250d3d1c-a665-4e77-8ac0-15768f6a25f1-kube-api-access-tfwf6\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.395004 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-combined-ca-bundle\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.395063 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-scripts\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.395107 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-config-data\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.395137 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfwf6\" (UniqueName: \"kubernetes.io/projected/250d3d1c-a665-4e77-8ac0-15768f6a25f1-kube-api-access-tfwf6\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.395185 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-fernet-keys\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.395260 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-credential-keys\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.401413 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-credential-keys\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.403763 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-combined-ca-bundle\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.406403 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a57822-e425-4ef7-8da8-f83a0c0abe60" path="/var/lib/kubelet/pods/60a57822-e425-4ef7-8da8-f83a0c0abe60/volumes" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.407200 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-config-data\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.409797 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-fernet-keys\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.410076 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-scripts\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.414139 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfwf6\" (UniqueName: \"kubernetes.io/projected/250d3d1c-a665-4e77-8ac0-15768f6a25f1-kube-api-access-tfwf6\") pod \"keystone-bootstrap-t8lkt\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.536223 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9flmk" Mar 09 18:47:19 crc kubenswrapper[4750]: I0309 18:47:19.543524 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:21 crc kubenswrapper[4750]: I0309 18:47:21.743952 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:47:21 crc kubenswrapper[4750]: I0309 18:47:21.744070 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:47:21 crc kubenswrapper[4750]: I0309 18:47:21.744155 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:47:21 crc kubenswrapper[4750]: I0309 18:47:21.745208 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aac3b01863d2a16c0edcf63137329635f54035a4884d4a4ea6dddfcacb5b31f0"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:47:21 crc kubenswrapper[4750]: I0309 18:47:21.745748 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://aac3b01863d2a16c0edcf63137329635f54035a4884d4a4ea6dddfcacb5b31f0" gracePeriod=600 Mar 09 18:47:21 crc kubenswrapper[4750]: I0309 18:47:21.960650 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Mar 09 18:47:22 crc kubenswrapper[4750]: I0309 18:47:22.745481 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="aac3b01863d2a16c0edcf63137329635f54035a4884d4a4ea6dddfcacb5b31f0" exitCode=0 Mar 09 18:47:22 crc kubenswrapper[4750]: I0309 18:47:22.745545 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"aac3b01863d2a16c0edcf63137329635f54035a4884d4a4ea6dddfcacb5b31f0"} Mar 09 18:47:22 crc kubenswrapper[4750]: I0309 18:47:22.745623 4750 scope.go:117] "RemoveContainer" containerID="998a8326795d648e01a6e195fe1e013eaedd975e8692a77e8f1a9a6dca44107f" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.654884 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.655252 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.655405 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n668hd9hd5h5b6h664h94h6dh8h5bbh96h5h5c9h574h549h68bh575h556h5c8h95h5c9h658h697h74h89h8dh677h79h697h9ch5cdh645h66cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gnpx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-85d485cb5-xkqdq_openstack(233d11ed-1edb-4e2d-b7b1-822685910821): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.658387 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-85d485cb5-xkqdq" podUID="233d11ed-1edb-4e2d-b7b1-822685910821" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.665824 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.665882 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.666021 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n564h5fdh74h5dbh57ch559h696h597h9dh78h689h5fbh67dh5cbh5f4h54h695hb4h67ch685h5dch574h57fh576h67chcch5fbh594h8bh5f5h698h698q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55x5z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6dcf6cd949-lbxvv_openstack(66e0e816-a509-43e6-824e-960e537d6bfc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.669059 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-6dcf6cd949-lbxvv" podUID="66e0e816-a509-43e6-824e-960e537d6bfc" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.675609 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.675674 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.675806 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc7h5c4h686h564h66bh5dfh657h65bhc8h654h547h5b5h558h6dhdfh78h579h668h685h5cdh4h579hcch649h98h5b9hc7h684h688h559h94h698q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ngq5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-d55fc75b5-pjhlj_openstack(bdb93bcb-fc2c-4480-85a4-709abaf41138): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:47:24 crc kubenswrapper[4750]: E0309 18:47:24.677799 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-d55fc75b5-pjhlj" podUID="bdb93bcb-fc2c-4480-85a4-709abaf41138" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.315915 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.316117 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.316246 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.110:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hbxrg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-59wvm_openstack(215dfd5c-6d06-42ef-89c7-6d40910fbf28): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.317454 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-59wvm" podUID="215dfd5c-6d06-42ef-89c7-6d40910fbf28" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.490157 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.494014 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650572 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-logs\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650623 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650681 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct6x4\" (UniqueName: \"kubernetes.io/projected/8573c870-7c80-4500-8e66-3ee2a6088b66-kube-api-access-ct6x4\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650811 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-combined-ca-bundle\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650851 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-scripts\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650886 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-combined-ca-bundle\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650909 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-public-tls-certs\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650950 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-internal-tls-certs\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.650983 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-logs\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.651030 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-httpd-run\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.651060 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-httpd-run\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.651094 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4nvl\" (UniqueName: \"kubernetes.io/projected/189e2dcb-e49f-4af8-8907-807a81f5dfd3-kube-api-access-p4nvl\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.651135 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-config-data\") pod \"8573c870-7c80-4500-8e66-3ee2a6088b66\" (UID: \"8573c870-7c80-4500-8e66-3ee2a6088b66\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.651180 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-config-data\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.651222 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-scripts\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.651279 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\" (UID: \"189e2dcb-e49f-4af8-8907-807a81f5dfd3\") " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.652236 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-logs" (OuterVolumeSpecName: "logs") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.652411 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-logs" (OuterVolumeSpecName: "logs") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.653110 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.653158 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.656233 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.656594 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.657949 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189e2dcb-e49f-4af8-8907-807a81f5dfd3-kube-api-access-p4nvl" (OuterVolumeSpecName: "kube-api-access-p4nvl") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "kube-api-access-p4nvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.658110 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-scripts" (OuterVolumeSpecName: "scripts") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.659440 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-scripts" (OuterVolumeSpecName: "scripts") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.680113 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.680382 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8573c870-7c80-4500-8e66-3ee2a6088b66-kube-api-access-ct6x4" (OuterVolumeSpecName: "kube-api-access-ct6x4") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "kube-api-access-ct6x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.683777 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.705327 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.706810 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-config-data" (OuterVolumeSpecName: "config-data") pod "189e2dcb-e49f-4af8-8907-807a81f5dfd3" (UID: "189e2dcb-e49f-4af8-8907-807a81f5dfd3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.713040 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.717279 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-config-data" (OuterVolumeSpecName: "config-data") pod "8573c870-7c80-4500-8e66-3ee2a6088b66" (UID: "8573c870-7c80-4500-8e66-3ee2a6088b66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752860 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752893 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752932 4750 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752944 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752962 4750 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752974 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct6x4\" (UniqueName: \"kubernetes.io/projected/8573c870-7c80-4500-8e66-3ee2a6088b66-kube-api-access-ct6x4\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752984 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.752996 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753015 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753026 4750 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753035 4750 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/189e2dcb-e49f-4af8-8907-807a81f5dfd3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753045 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753056 4750 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/189e2dcb-e49f-4af8-8907-807a81f5dfd3-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753067 4750 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8573c870-7c80-4500-8e66-3ee2a6088b66-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753077 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4nvl\" (UniqueName: \"kubernetes.io/projected/189e2dcb-e49f-4af8-8907-807a81f5dfd3-kube-api-access-p4nvl\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.753090 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8573c870-7c80-4500-8e66-3ee2a6088b66-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.777103 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8573c870-7c80-4500-8e66-3ee2a6088b66","Type":"ContainerDied","Data":"1288f158988b1e47e72a991e7fb035688a93eef51e57e58d7a327f0a4248d135"} Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.777217 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.777570 4750 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.779669 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"189e2dcb-e49f-4af8-8907-807a81f5dfd3","Type":"ContainerDied","Data":"317bd1028df012ad0e9762fdc70b7b94a7cf5239ca6c6ff25dc17b201ac4f7f8"} Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.779778 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.781909 4750 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.782757 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-59wvm" podUID="215dfd5c-6d06-42ef-89c7-6d40910fbf28" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.854657 4750 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.854701 4750 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.857058 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.867597 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.878180 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.887734 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.896039 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.896730 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-httpd" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.896745 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-httpd" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.896767 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-log" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.896774 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-log" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.896820 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-log" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.896828 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-log" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.896847 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-httpd" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.896853 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-httpd" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.897016 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-httpd" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.897029 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-httpd" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.897039 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" containerName="glance-log" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.897055 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" containerName="glance-log" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.898075 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.900968 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.901727 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.901955 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.902673 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qbg7p" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.905413 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.915947 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.917817 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.924412 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.926431 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 09 18:47:25 crc kubenswrapper[4750]: I0309 18:47:25.926704 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.992458 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.992510 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Mar 09 18:47:25 crc kubenswrapper[4750]: E0309 18:47:25.992678 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.102.83.110:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n94hb7h5f8h54fh5f9h5cbhd6h5dfhb7h5b8h564h66fhc8h67ch647h5ddhcfh574h9h5d6h9bh69h5dbh657h685hf7hf5h644h59h55bh74h79q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-64gbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(4d6c59a9-358a-4f97-8c95-f1a23f3854be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058414 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058473 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058498 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-logs\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058526 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058554 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058589 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t87kn\" (UniqueName: \"kubernetes.io/projected/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-kube-api-access-t87kn\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058614 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058685 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j65hg\" (UniqueName: \"kubernetes.io/projected/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-kube-api-access-j65hg\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058710 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058753 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-logs\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058781 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058810 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.058834 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.059090 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.059328 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.059401 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161257 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161317 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-logs\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161367 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161392 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161424 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t87kn\" (UniqueName: \"kubernetes.io/projected/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-kube-api-access-t87kn\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161452 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161492 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j65hg\" (UniqueName: \"kubernetes.io/projected/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-kube-api-access-j65hg\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161519 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161560 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-logs\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161589 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161618 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161659 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161691 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161763 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161796 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.161860 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.162010 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.162018 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.162133 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-logs\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.162319 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.162462 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-logs\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.166962 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.167076 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.167155 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.168866 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.171903 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.181792 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.186668 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t87kn\" (UniqueName: \"kubernetes.io/projected/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-kube-api-access-t87kn\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.190330 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.190917 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.202563 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j65hg\" (UniqueName: \"kubernetes.io/projected/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-kube-api-access-j65hg\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.224370 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.224948 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.242087 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:47:26 crc kubenswrapper[4750]: I0309 18:47:26.517824 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:27 crc kubenswrapper[4750]: I0309 18:47:27.396328 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="189e2dcb-e49f-4af8-8907-807a81f5dfd3" path="/var/lib/kubelet/pods/189e2dcb-e49f-4af8-8907-807a81f5dfd3/volumes" Mar 09 18:47:27 crc kubenswrapper[4750]: I0309 18:47:27.398948 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8573c870-7c80-4500-8e66-3ee2a6088b66" path="/var/lib/kubelet/pods/8573c870-7c80-4500-8e66-3ee2a6088b66/volumes" Mar 09 18:47:30 crc kubenswrapper[4750]: I0309 18:47:30.836805 4750 generic.go:334] "Generic (PLEG): container finished" podID="620fb043-252f-4bdc-8120-68a3978e1c82" containerID="79d793d41819e39872483071d0df21fac11e13760912f8b4ee73535605ed9d8a" exitCode=0 Mar 09 18:47:30 crc kubenswrapper[4750]: I0309 18:47:30.836879 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-x6m2l" event={"ID":"620fb043-252f-4bdc-8120-68a3978e1c82","Type":"ContainerDied","Data":"79d793d41819e39872483071d0df21fac11e13760912f8b4ee73535605ed9d8a"} Mar 09 18:47:31 crc kubenswrapper[4750]: I0309 18:47:31.961467 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Mar 09 18:47:31 crc kubenswrapper[4750]: I0309 18:47:31.961715 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.910449 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dcf6cd949-lbxvv" event={"ID":"66e0e816-a509-43e6-824e-960e537d6bfc","Type":"ContainerDied","Data":"3c57aa10969e8c7a8a2febcb1460bbaa9e3895235aa005486f7577e2222b0075"} Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.911227 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c57aa10969e8c7a8a2febcb1460bbaa9e3895235aa005486f7577e2222b0075" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.913559 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d55fc75b5-pjhlj" event={"ID":"bdb93bcb-fc2c-4480-85a4-709abaf41138","Type":"ContainerDied","Data":"404bc2c3426daa156c2a7df3e8d29207f6b67aef7935fe69cbe4e298e98c99c7"} Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.913620 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="404bc2c3426daa156c2a7df3e8d29207f6b67aef7935fe69cbe4e298e98c99c7" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.916738 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" event={"ID":"66f23451-0dce-4bfa-95e7-8d17adc315d0","Type":"ContainerDied","Data":"b81a8455253a9b8b62c93750d5fcc0be564f9fd1b2dc7974773e0b02e31a8a2a"} Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.916784 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b81a8455253a9b8b62c93750d5fcc0be564f9fd1b2dc7974773e0b02e31a8a2a" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.919278 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-x6m2l" event={"ID":"620fb043-252f-4bdc-8120-68a3978e1c82","Type":"ContainerDied","Data":"0d71fc431522ae1417fec7191123adf4c3004de0bde72099946fcc8c45aecb27"} Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.919320 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d71fc431522ae1417fec7191123adf4c3004de0bde72099946fcc8c45aecb27" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.921107 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d485cb5-xkqdq" event={"ID":"233d11ed-1edb-4e2d-b7b1-822685910821","Type":"ContainerDied","Data":"024a6808b63b4f55752324613cf87d6011eb796cd6dd0b519854853fd6216488"} Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.921144 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="024a6808b63b4f55752324613cf87d6011eb796cd6dd0b519854853fd6216488" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.963130 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.969461 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.977928 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.989893 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:47:35 crc kubenswrapper[4750]: I0309 18:47:35.993764 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.085323 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngq5b\" (UniqueName: \"kubernetes.io/projected/bdb93bcb-fc2c-4480-85a4-709abaf41138-kube-api-access-ngq5b\") pod \"bdb93bcb-fc2c-4480-85a4-709abaf41138\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.085390 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdb93bcb-fc2c-4480-85a4-709abaf41138-horizon-secret-key\") pod \"bdb93bcb-fc2c-4480-85a4-709abaf41138\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.085446 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-scripts\") pod \"66e0e816-a509-43e6-824e-960e537d6bfc\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.085837 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4vgf\" (UniqueName: \"kubernetes.io/projected/620fb043-252f-4bdc-8120-68a3978e1c82-kube-api-access-n4vgf\") pod \"620fb043-252f-4bdc-8120-68a3978e1c82\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.085872 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-scripts\") pod \"bdb93bcb-fc2c-4480-85a4-709abaf41138\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.085946 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233d11ed-1edb-4e2d-b7b1-822685910821-logs\") pod \"233d11ed-1edb-4e2d-b7b1-822685910821\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.085997 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66e0e816-a509-43e6-824e-960e537d6bfc-logs\") pod \"66e0e816-a509-43e6-824e-960e537d6bfc\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086020 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-nb\") pod \"66f23451-0dce-4bfa-95e7-8d17adc315d0\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086083 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55x5z\" (UniqueName: \"kubernetes.io/projected/66e0e816-a509-43e6-824e-960e537d6bfc-kube-api-access-55x5z\") pod \"66e0e816-a509-43e6-824e-960e537d6bfc\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086102 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-combined-ca-bundle\") pod \"620fb043-252f-4bdc-8120-68a3978e1c82\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086158 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-config-data\") pod \"bdb93bcb-fc2c-4480-85a4-709abaf41138\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086180 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-config-data\") pod \"233d11ed-1edb-4e2d-b7b1-822685910821\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086205 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-config-data\") pod \"66e0e816-a509-43e6-824e-960e537d6bfc\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086273 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnpx8\" (UniqueName: \"kubernetes.io/projected/233d11ed-1edb-4e2d-b7b1-822685910821-kube-api-access-gnpx8\") pod \"233d11ed-1edb-4e2d-b7b1-822685910821\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086308 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233d11ed-1edb-4e2d-b7b1-822685910821-horizon-secret-key\") pod \"233d11ed-1edb-4e2d-b7b1-822685910821\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086359 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-scripts\") pod \"233d11ed-1edb-4e2d-b7b1-822685910821\" (UID: \"233d11ed-1edb-4e2d-b7b1-822685910821\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086240 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-scripts" (OuterVolumeSpecName: "scripts") pod "66e0e816-a509-43e6-824e-960e537d6bfc" (UID: "66e0e816-a509-43e6-824e-960e537d6bfc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086404 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-config\") pod \"620fb043-252f-4bdc-8120-68a3978e1c82\" (UID: \"620fb043-252f-4bdc-8120-68a3978e1c82\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086455 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-scripts" (OuterVolumeSpecName: "scripts") pod "bdb93bcb-fc2c-4480-85a4-709abaf41138" (UID: "bdb93bcb-fc2c-4480-85a4-709abaf41138"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086546 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-config\") pod \"66f23451-0dce-4bfa-95e7-8d17adc315d0\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086655 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-sb\") pod \"66f23451-0dce-4bfa-95e7-8d17adc315d0\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086725 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb93bcb-fc2c-4480-85a4-709abaf41138-logs\") pod \"bdb93bcb-fc2c-4480-85a4-709abaf41138\" (UID: \"bdb93bcb-fc2c-4480-85a4-709abaf41138\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086765 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66e0e816-a509-43e6-824e-960e537d6bfc-horizon-secret-key\") pod \"66e0e816-a509-43e6-824e-960e537d6bfc\" (UID: \"66e0e816-a509-43e6-824e-960e537d6bfc\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.086798 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-dns-svc\") pod \"66f23451-0dce-4bfa-95e7-8d17adc315d0\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.087168 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-config-data" (OuterVolumeSpecName: "config-data") pod "bdb93bcb-fc2c-4480-85a4-709abaf41138" (UID: "bdb93bcb-fc2c-4480-85a4-709abaf41138"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.087817 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-config-data" (OuterVolumeSpecName: "config-data") pod "233d11ed-1edb-4e2d-b7b1-822685910821" (UID: "233d11ed-1edb-4e2d-b7b1-822685910821"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.087870 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/233d11ed-1edb-4e2d-b7b1-822685910821-logs" (OuterVolumeSpecName: "logs") pod "233d11ed-1edb-4e2d-b7b1-822685910821" (UID: "233d11ed-1edb-4e2d-b7b1-822685910821"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.087926 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-config-data" (OuterVolumeSpecName: "config-data") pod "66e0e816-a509-43e6-824e-960e537d6bfc" (UID: "66e0e816-a509-43e6-824e-960e537d6bfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.087978 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdb93bcb-fc2c-4480-85a4-709abaf41138-logs" (OuterVolumeSpecName: "logs") pod "bdb93bcb-fc2c-4480-85a4-709abaf41138" (UID: "bdb93bcb-fc2c-4480-85a4-709abaf41138"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.088200 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66e0e816-a509-43e6-824e-960e537d6bfc-logs" (OuterVolumeSpecName: "logs") pod "66e0e816-a509-43e6-824e-960e537d6bfc" (UID: "66e0e816-a509-43e6-824e-960e537d6bfc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.088777 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.088807 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.088820 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdb93bcb-fc2c-4480-85a4-709abaf41138-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.104983 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdb93bcb-fc2c-4480-85a4-709abaf41138-kube-api-access-ngq5b" (OuterVolumeSpecName: "kube-api-access-ngq5b") pod "bdb93bcb-fc2c-4480-85a4-709abaf41138" (UID: "bdb93bcb-fc2c-4480-85a4-709abaf41138"). InnerVolumeSpecName "kube-api-access-ngq5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.105385 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-scripts" (OuterVolumeSpecName: "scripts") pod "233d11ed-1edb-4e2d-b7b1-822685910821" (UID: "233d11ed-1edb-4e2d-b7b1-822685910821"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.107087 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb93bcb-fc2c-4480-85a4-709abaf41138-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bdb93bcb-fc2c-4480-85a4-709abaf41138" (UID: "bdb93bcb-fc2c-4480-85a4-709abaf41138"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.107099 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/620fb043-252f-4bdc-8120-68a3978e1c82-kube-api-access-n4vgf" (OuterVolumeSpecName: "kube-api-access-n4vgf") pod "620fb043-252f-4bdc-8120-68a3978e1c82" (UID: "620fb043-252f-4bdc-8120-68a3978e1c82"). InnerVolumeSpecName "kube-api-access-n4vgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.107249 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66e0e816-a509-43e6-824e-960e537d6bfc-kube-api-access-55x5z" (OuterVolumeSpecName: "kube-api-access-55x5z") pod "66e0e816-a509-43e6-824e-960e537d6bfc" (UID: "66e0e816-a509-43e6-824e-960e537d6bfc"). InnerVolumeSpecName "kube-api-access-55x5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.108090 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/233d11ed-1edb-4e2d-b7b1-822685910821-kube-api-access-gnpx8" (OuterVolumeSpecName: "kube-api-access-gnpx8") pod "233d11ed-1edb-4e2d-b7b1-822685910821" (UID: "233d11ed-1edb-4e2d-b7b1-822685910821"). InnerVolumeSpecName "kube-api-access-gnpx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.108819 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/233d11ed-1edb-4e2d-b7b1-822685910821-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "233d11ed-1edb-4e2d-b7b1-822685910821" (UID: "233d11ed-1edb-4e2d-b7b1-822685910821"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.110351 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e0e816-a509-43e6-824e-960e537d6bfc-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "66e0e816-a509-43e6-824e-960e537d6bfc" (UID: "66e0e816-a509-43e6-824e-960e537d6bfc"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.124666 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "620fb043-252f-4bdc-8120-68a3978e1c82" (UID: "620fb043-252f-4bdc-8120-68a3978e1c82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.142981 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "66f23451-0dce-4bfa-95e7-8d17adc315d0" (UID: "66f23451-0dce-4bfa-95e7-8d17adc315d0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.147819 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-config" (OuterVolumeSpecName: "config") pod "620fb043-252f-4bdc-8120-68a3978e1c82" (UID: "620fb043-252f-4bdc-8120-68a3978e1c82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.158240 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "66f23451-0dce-4bfa-95e7-8d17adc315d0" (UID: "66f23451-0dce-4bfa-95e7-8d17adc315d0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.161796 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "66f23451-0dce-4bfa-95e7-8d17adc315d0" (UID: "66f23451-0dce-4bfa-95e7-8d17adc315d0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.172526 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-config" (OuterVolumeSpecName: "config") pod "66f23451-0dce-4bfa-95e7-8d17adc315d0" (UID: "66f23451-0dce-4bfa-95e7-8d17adc315d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.190591 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s42r\" (UniqueName: \"kubernetes.io/projected/66f23451-0dce-4bfa-95e7-8d17adc315d0-kube-api-access-2s42r\") pod \"66f23451-0dce-4bfa-95e7-8d17adc315d0\" (UID: \"66f23451-0dce-4bfa-95e7-8d17adc315d0\") " Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191624 4750 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233d11ed-1edb-4e2d-b7b1-822685910821-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191656 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191667 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191676 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191684 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191693 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb93bcb-fc2c-4480-85a4-709abaf41138-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191705 4750 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66e0e816-a509-43e6-824e-960e537d6bfc-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191713 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191722 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngq5b\" (UniqueName: \"kubernetes.io/projected/bdb93bcb-fc2c-4480-85a4-709abaf41138-kube-api-access-ngq5b\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191732 4750 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdb93bcb-fc2c-4480-85a4-709abaf41138-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191743 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4vgf\" (UniqueName: \"kubernetes.io/projected/620fb043-252f-4bdc-8120-68a3978e1c82-kube-api-access-n4vgf\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191752 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233d11ed-1edb-4e2d-b7b1-822685910821-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191761 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66e0e816-a509-43e6-824e-960e537d6bfc-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191773 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66f23451-0dce-4bfa-95e7-8d17adc315d0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191783 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55x5z\" (UniqueName: \"kubernetes.io/projected/66e0e816-a509-43e6-824e-960e537d6bfc-kube-api-access-55x5z\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191791 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fb043-252f-4bdc-8120-68a3978e1c82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191800 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233d11ed-1edb-4e2d-b7b1-822685910821-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191808 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66e0e816-a509-43e6-824e-960e537d6bfc-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.191819 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnpx8\" (UniqueName: \"kubernetes.io/projected/233d11ed-1edb-4e2d-b7b1-822685910821-kube-api-access-gnpx8\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.195287 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66f23451-0dce-4bfa-95e7-8d17adc315d0-kube-api-access-2s42r" (OuterVolumeSpecName: "kube-api-access-2s42r") pod "66f23451-0dce-4bfa-95e7-8d17adc315d0" (UID: "66f23451-0dce-4bfa-95e7-8d17adc315d0"). InnerVolumeSpecName "kube-api-access-2s42r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.292572 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s42r\" (UniqueName: \"kubernetes.io/projected/66f23451-0dce-4bfa-95e7-8d17adc315d0-kube-api-access-2s42r\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.928928 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d485cb5-xkqdq" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.928995 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.929013 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-x6m2l" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.928995 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dcf6cd949-lbxvv" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.928996 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d55fc75b5-pjhlj" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.962422 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cd9f6b887-2nlmb" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Mar 09 18:47:36 crc kubenswrapper[4750]: I0309 18:47:36.994078 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cd9f6b887-2nlmb"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.023920 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cd9f6b887-2nlmb"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.062001 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85d485cb5-xkqdq"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.090982 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85d485cb5-xkqdq"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.114251 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6dcf6cd949-lbxvv"] Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.121184 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.121253 4750 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.110:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.121399 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.110:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-chvrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-89k9z_openstack(fec39559-ddae-4e19-8941-0fc3209084cd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.122871 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-89k9z" podUID="fec39559-ddae-4e19-8941-0fc3209084cd" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.130068 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6dcf6cd949-lbxvv"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.182754 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d55fc75b5-pjhlj"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.228166 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d55fc75b5-pjhlj"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.292406 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb69d7559-cxkgc"] Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.292820 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="620fb043-252f-4bdc-8120-68a3978e1c82" containerName="neutron-db-sync" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.292831 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="620fb043-252f-4bdc-8120-68a3978e1c82" containerName="neutron-db-sync" Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.292850 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.292856 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.292869 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="init" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.292875 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="init" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.293035 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" containerName="dnsmasq-dns" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.293051 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="620fb043-252f-4bdc-8120-68a3978e1c82" containerName="neutron-db-sync" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.294136 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.314069 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb69d7559-cxkgc"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.405109 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="233d11ed-1edb-4e2d-b7b1-822685910821" path="/var/lib/kubelet/pods/233d11ed-1edb-4e2d-b7b1-822685910821/volumes" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.406085 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66e0e816-a509-43e6-824e-960e537d6bfc" path="/var/lib/kubelet/pods/66e0e816-a509-43e6-824e-960e537d6bfc/volumes" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.406516 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66f23451-0dce-4bfa-95e7-8d17adc315d0" path="/var/lib/kubelet/pods/66f23451-0dce-4bfa-95e7-8d17adc315d0/volumes" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.408266 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdb93bcb-fc2c-4480-85a4-709abaf41138" path="/var/lib/kubelet/pods/bdb93bcb-fc2c-4480-85a4-709abaf41138/volumes" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.409313 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d6db5794d-f9pxj"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.411419 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.414071 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.414615 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.416520 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.418602 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d6db5794d-f9pxj"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.418775 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wklpd\" (UniqueName: \"kubernetes.io/projected/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-kube-api-access-wklpd\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.418863 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.418911 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-nb\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.419140 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4p74c" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.419227 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-svc\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.419372 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-config\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.419425 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-swift-storage-0\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.520987 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-config\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521040 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-swift-storage-0\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521219 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wklpd\" (UniqueName: \"kubernetes.io/projected/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-kube-api-access-wklpd\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521308 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521330 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-nb\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521468 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8rz4\" (UniqueName: \"kubernetes.io/projected/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-kube-api-access-z8rz4\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521521 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-ovndb-tls-certs\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521555 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-svc\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521582 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-config\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.521618 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-combined-ca-bundle\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.523026 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-svc\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.523066 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-nb\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.523695 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.524030 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-httpd-config\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.525450 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-swift-storage-0\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.526585 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-config\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.547516 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wklpd\" (UniqueName: \"kubernetes.io/projected/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-kube-api-access-wklpd\") pod \"dnsmasq-dns-7cb69d7559-cxkgc\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.559522 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fdd5dd964-x7n8q"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.627667 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8rz4\" (UniqueName: \"kubernetes.io/projected/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-kube-api-access-z8rz4\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.627711 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-ovndb-tls-certs\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.627746 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-config\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.627764 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-combined-ca-bundle\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.627791 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-httpd-config\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.631724 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-httpd-config\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.637305 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-ovndb-tls-certs\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.637378 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ddc4b4b8d-f5q7t"] Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.641041 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-config\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.643421 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-combined-ca-bundle\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.644312 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.648638 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8rz4\" (UniqueName: \"kubernetes.io/projected/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-kube-api-access-z8rz4\") pod \"neutron-5d6db5794d-f9pxj\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.787111 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.950624 4750 generic.go:334] "Generic (PLEG): container finished" podID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerID="772d7a0f90ac442db02b04642c9d04d45eef711be2787a15008d14446a3959ee" exitCode=137 Mar 09 18:47:37 crc kubenswrapper[4750]: I0309 18:47:37.950685 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433","Type":"ContainerDied","Data":"772d7a0f90ac442db02b04642c9d04d45eef711be2787a15008d14446a3959ee"} Mar 09 18:47:37 crc kubenswrapper[4750]: E0309 18:47:37.952608 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.110:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-89k9z" podUID="fec39559-ddae-4e19-8941-0fc3209084cd" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.124221 4750 scope.go:117] "RemoveContainer" containerID="e721b26c39b4bd7db5346d1e67ec0fade4b05f0b0caaad98c67c833c88f5ca4f" Mar 09 18:47:38 crc kubenswrapper[4750]: W0309 18:47:38.172442 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef45ec45_1ef2_46e6_bf33_0c092885679f.slice/crio-47c7d7888f925219c0f83327043d08e690f53d43cb175c3af95161e414de0531 WatchSource:0}: Error finding container 47c7d7888f925219c0f83327043d08e690f53d43cb175c3af95161e414de0531: Status 404 returned error can't find the container with id 47c7d7888f925219c0f83327043d08e690f53d43cb175c3af95161e414de0531 Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.377054 4750 scope.go:117] "RemoveContainer" containerID="bf0754d343051a78eff0c4d13faffa6e7f84a2ad8cae7fc86bae0da0abc4156a" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.480863 4750 scope.go:117] "RemoveContainer" containerID="5ea9b1043fc3d3c485d640964e713fd62560bdcf4e3e7584547370dd29b722de" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.617843 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t8lkt"] Mar 09 18:47:38 crc kubenswrapper[4750]: W0309 18:47:38.638830 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod250d3d1c_a665_4e77_8ac0_15768f6a25f1.slice/crio-2bfc3ac1b91b7658d08225b83ff4b7111fc8a4daaa14e54bc8ec8944f062338a WatchSource:0}: Error finding container 2bfc3ac1b91b7658d08225b83ff4b7111fc8a4daaa14e54bc8ec8944f062338a: Status 404 returned error can't find the container with id 2bfc3ac1b91b7658d08225b83ff4b7111fc8a4daaa14e54bc8ec8944f062338a Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.679145 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.695453 4750 scope.go:117] "RemoveContainer" containerID="42c659a821382be1ef3d4981241e21783ff08f6a4ea3b8ac936124dafe96e0ac" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.711792 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.861427 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-logs\") pod \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.861478 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-custom-prometheus-ca\") pod \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.861588 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-combined-ca-bundle\") pod \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.861639 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh7tw\" (UniqueName: \"kubernetes.io/projected/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-kube-api-access-lh7tw\") pod \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.861771 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-config-data\") pod \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\" (UID: \"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433\") " Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.864907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-logs" (OuterVolumeSpecName: "logs") pod "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" (UID: "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.879421 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-kube-api-access-lh7tw" (OuterVolumeSpecName: "kube-api-access-lh7tw") pod "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" (UID: "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433"). InnerVolumeSpecName "kube-api-access-lh7tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.950259 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" (UID: "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.961866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"53dae358-d7dd-4823-8369-d548e214333e","Type":"ContainerStarted","Data":"403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d"} Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.963609 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t8lkt" event={"ID":"250d3d1c-a665-4e77-8ac0-15768f6a25f1","Type":"ContainerStarted","Data":"2bfc3ac1b91b7658d08225b83ff4b7111fc8a4daaa14e54bc8ec8944f062338a"} Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.963689 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.963705 4750 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.963714 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh7tw\" (UniqueName: \"kubernetes.io/projected/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-kube-api-access-lh7tw\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.965933 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"0bf642275aba662d4b9920780da985536f4c7bcb2a54a88c1149087655cecbbd"} Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.967063 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fdd5dd964-x7n8q" event={"ID":"0509f0f1-226d-4626-998d-b9065a8a634b","Type":"ContainerStarted","Data":"6b9396e4cc10e6f66f53eaa89c38338f7c33763087615bbc9f12a5a97ea807b1"} Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.970443 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.970496 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433","Type":"ContainerDied","Data":"ba7153ca4c17724d4bdfc1dbb6a92741a5124afee90649f64118c94683a4c84c"} Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.970575 4750 scope.go:117] "RemoveContainer" containerID="772d7a0f90ac442db02b04642c9d04d45eef711be2787a15008d14446a3959ee" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.981625 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddc4b4b8d-f5q7t" event={"ID":"ef45ec45-1ef2-46e6-bf33-0c092885679f","Type":"ContainerStarted","Data":"47c7d7888f925219c0f83327043d08e690f53d43cb175c3af95161e414de0531"} Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.982222 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=16.582265476 podStartE2EDuration="35.982200838s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="2026-03-09 18:47:05.908471904 +0000 UTC m=+1307.250944302" lastFinishedPulling="2026-03-09 18:47:25.308407266 +0000 UTC m=+1326.650879664" observedRunningTime="2026-03-09 18:47:38.978301082 +0000 UTC m=+1340.320773480" watchObservedRunningTime="2026-03-09 18:47:38.982200838 +0000 UTC m=+1340.324673236" Mar 09 18:47:38 crc kubenswrapper[4750]: I0309 18:47:38.994336 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsgm7" event={"ID":"2afe8078-9724-4ab3-b5ee-79c978ce9bea","Type":"ContainerStarted","Data":"8e3d7cd0c864c75c3c552a0a15ab5245d401dcf5c5c8c397a62a26b48376969d"} Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.003471 4750 scope.go:117] "RemoveContainer" containerID="40bdcea7bb85288f1c4bbb741b609100cb12caba25eee36bab891833ab9fec41" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.008675 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:47:39 crc kubenswrapper[4750]: W0309 18:47:39.022422 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd177b683_bf5e_4048_a0cf_2b7e6176f2ed.slice/crio-6a2c6af7749343b6d46928b0042c01d49f80d01d14b2d99b0c02e006736f62c0 WatchSource:0}: Error finding container 6a2c6af7749343b6d46928b0042c01d49f80d01d14b2d99b0c02e006736f62c0: Status 404 returned error can't find the container with id 6a2c6af7749343b6d46928b0042c01d49f80d01d14b2d99b0c02e006736f62c0 Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.022581 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"9d32a26f-bd14-41a7-872b-e102ea95f696","Type":"ContainerStarted","Data":"793969d3d3e5cab9242146e23eb4c826c91cafe7e607d142e01d40eb07324c7c"} Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.029636 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-hsgm7" podStartSLOduration=5.5575843240000005 podStartE2EDuration="36.029613065s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="2026-03-09 18:47:06.576642416 +0000 UTC m=+1307.919114814" lastFinishedPulling="2026-03-09 18:47:37.048671157 +0000 UTC m=+1338.391143555" observedRunningTime="2026-03-09 18:47:39.026973444 +0000 UTC m=+1340.369445842" watchObservedRunningTime="2026-03-09 18:47:39.029613065 +0000 UTC m=+1340.372085453" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.051464 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=16.288587577 podStartE2EDuration="36.051438269s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="2026-03-09 18:47:05.568407315 +0000 UTC m=+1306.910879713" lastFinishedPulling="2026-03-09 18:47:25.331258007 +0000 UTC m=+1326.673730405" observedRunningTime="2026-03-09 18:47:39.046531896 +0000 UTC m=+1340.389004294" watchObservedRunningTime="2026-03-09 18:47:39.051438269 +0000 UTC m=+1340.393910667" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.134957 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb69d7559-cxkgc"] Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.144567 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-config-data" (OuterVolumeSpecName: "config-data") pod "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" (UID: "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.154568 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" (UID: "3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.170047 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.170083 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.262091 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d6db5794d-f9pxj"] Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.571319 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Mar 09 18:47:39 crc kubenswrapper[4750]: I0309 18:47:39.804699 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:47:39 crc kubenswrapper[4750]: W0309 18:47:39.816860 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91ab0cb5_95bc_4e78_9acc_9a0414c6c25e.slice/crio-d618d2e49e51c911395fe859fc17574fe7574c814eb40f6e5cb5962957e8062e WatchSource:0}: Error finding container d618d2e49e51c911395fe859fc17574fe7574c814eb40f6e5cb5962957e8062e: Status 404 returned error can't find the container with id d618d2e49e51c911395fe859fc17574fe7574c814eb40f6e5cb5962957e8062e Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.046486 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-748768969-j7m74"] Mar 09 18:47:40 crc kubenswrapper[4750]: E0309 18:47:40.048352 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api-log" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.048370 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api-log" Mar 09 18:47:40 crc kubenswrapper[4750]: E0309 18:47:40.048396 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.048402 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.048557 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api-log" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.048584 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" containerName="watcher-api" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.049580 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.066416 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.067261 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.092733 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-748768969-j7m74"] Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.113194 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" event={"ID":"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e","Type":"ContainerStarted","Data":"b8c0fc0c526c386f4ceaab293fe2b74773add67bcb9853549198cf3b69c98ca6"} Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.208857 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d177b683-bf5e-4048-a0cf-2b7e6176f2ed","Type":"ContainerStarted","Data":"6a2c6af7749343b6d46928b0042c01d49f80d01d14b2d99b0c02e006736f62c0"} Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.210760 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-httpd-config\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.210921 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-combined-ca-bundle\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.210961 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfhxl\" (UniqueName: \"kubernetes.io/projected/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-kube-api-access-jfhxl\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.210992 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-public-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.211047 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-internal-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.211219 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-config\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.211286 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-ovndb-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.215265 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddc4b4b8d-f5q7t" event={"ID":"ef45ec45-1ef2-46e6-bf33-0c092885679f","Type":"ContainerStarted","Data":"7c882d7feb763fa768bb968ae79a038f49e22c03454de50f1ddfe39b393e082b"} Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.218161 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e","Type":"ContainerStarted","Data":"d618d2e49e51c911395fe859fc17574fe7574c814eb40f6e5cb5962957e8062e"} Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.222121 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerStarted","Data":"6e6034f7d85033a58557e0e8af1f961c2ebd2a12c6f963ee1aa5060d9ca53b8e"} Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.256847 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t8lkt" event={"ID":"250d3d1c-a665-4e77-8ac0-15768f6a25f1","Type":"ContainerStarted","Data":"7ccb2dde1714b67016d32aa2bf42313d9a79e120ae7ad1e8ed5ac4e14697f4ea"} Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.316080 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-internal-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.316197 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-config\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.316236 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-ovndb-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.316295 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-httpd-config\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.316344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-combined-ca-bundle\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.316368 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfhxl\" (UniqueName: \"kubernetes.io/projected/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-kube-api-access-jfhxl\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.316390 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-public-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.332979 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6db5794d-f9pxj" event={"ID":"9d5b4b2b-6058-44ac-b24b-e5ef2423f830","Type":"ContainerStarted","Data":"779e4b67cda39bfa39b4708149b50445e378fd8f66e9116a49fc6b8097ea96b5"} Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.341090 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-public-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.343829 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-config\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.344369 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-httpd-config\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.344959 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-internal-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.345096 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-combined-ca-bundle\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.355162 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-ovndb-tls-certs\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.371596 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfhxl\" (UniqueName: \"kubernetes.io/projected/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-kube-api-access-jfhxl\") pod \"neutron-748768969-j7m74\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: I0309 18:47:40.415486 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:40 crc kubenswrapper[4750]: E0309 18:47:40.549938 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc55e5e2c_4e0b_4d7b_85e6_9cb49c96993e.slice/crio-conmon-0320c6c25a7f5817ddc2e5b13113e5077f79155cbc555208882b4c19bb3fbedf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc55e5e2c_4e0b_4d7b_85e6_9cb49c96993e.slice/crio-0320c6c25a7f5817ddc2e5b13113e5077f79155cbc555208882b4c19bb3fbedf.scope\": RecentStats: unable to find data in memory cache]" Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.076851 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t8lkt" podStartSLOduration=22.076832895 podStartE2EDuration="22.076832895s" podCreationTimestamp="2026-03-09 18:47:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:40.319322745 +0000 UTC m=+1341.661795163" watchObservedRunningTime="2026-03-09 18:47:41.076832895 +0000 UTC m=+1342.419305293" Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.085141 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-748768969-j7m74"] Mar 09 18:47:41 crc kubenswrapper[4750]: W0309 18:47:41.107299 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72dade1d_7fb7_4118_a665_7b2fdfc56a9e.slice/crio-84c5cffc803cc82228e944e18adf5d58dbee69b2e90b171697d30a95e48b5917 WatchSource:0}: Error finding container 84c5cffc803cc82228e944e18adf5d58dbee69b2e90b171697d30a95e48b5917: Status 404 returned error can't find the container with id 84c5cffc803cc82228e944e18adf5d58dbee69b2e90b171697d30a95e48b5917 Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.449727 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fdd5dd964-x7n8q" event={"ID":"0509f0f1-226d-4626-998d-b9065a8a634b","Type":"ContainerStarted","Data":"6ad27ced973934752e2b2043e32063300f415f2da353a61711c5f28af86eb152"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.450088 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fdd5dd964-x7n8q" event={"ID":"0509f0f1-226d-4626-998d-b9065a8a634b","Type":"ContainerStarted","Data":"30968d93cb1e9dd5e9b87ad6c7b0c20b0f208e717272ca2bf175dd7bc9c6e2a5"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.450098 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748768969-j7m74" event={"ID":"72dade1d-7fb7-4118-a665-7b2fdfc56a9e","Type":"ContainerStarted","Data":"84c5cffc803cc82228e944e18adf5d58dbee69b2e90b171697d30a95e48b5917"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.458761 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6db5794d-f9pxj" event={"ID":"9d5b4b2b-6058-44ac-b24b-e5ef2423f830","Type":"ContainerStarted","Data":"5ad8d0972171dc5644b0c875da08f240816d847855e0e2602a5e8575930a3e3e"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.458845 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6db5794d-f9pxj" event={"ID":"9d5b4b2b-6058-44ac-b24b-e5ef2423f830","Type":"ContainerStarted","Data":"f9d8d985d34814bb2928cc6886169310c947dcea0cf7abd0d8398aa3e3a01f7c"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.460159 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.476494 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7fdd5dd964-x7n8q" podStartSLOduration=27.949957639 podStartE2EDuration="28.476471893s" podCreationTimestamp="2026-03-09 18:47:13 +0000 UTC" firstStartedPulling="2026-03-09 18:47:38.180041205 +0000 UTC m=+1339.522513603" lastFinishedPulling="2026-03-09 18:47:38.706555459 +0000 UTC m=+1340.049027857" observedRunningTime="2026-03-09 18:47:41.450050385 +0000 UTC m=+1342.792522783" watchObservedRunningTime="2026-03-09 18:47:41.476471893 +0000 UTC m=+1342.818944291" Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.488712 4750 generic.go:334] "Generic (PLEG): container finished" podID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerID="0320c6c25a7f5817ddc2e5b13113e5077f79155cbc555208882b4c19bb3fbedf" exitCode=0 Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.488856 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" event={"ID":"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e","Type":"ContainerDied","Data":"0320c6c25a7f5817ddc2e5b13113e5077f79155cbc555208882b4c19bb3fbedf"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.489826 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d6db5794d-f9pxj" podStartSLOduration=4.489807015 podStartE2EDuration="4.489807015s" podCreationTimestamp="2026-03-09 18:47:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:41.482742063 +0000 UTC m=+1342.825214481" watchObservedRunningTime="2026-03-09 18:47:41.489807015 +0000 UTC m=+1342.832279413" Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.523184 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddc4b4b8d-f5q7t" event={"ID":"ef45ec45-1ef2-46e6-bf33-0c092885679f","Type":"ContainerStarted","Data":"8f715bad4ddba91fed463c55ef1bb23df16777368f645f8c1a8928028b49eb87"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.573711 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-59wvm" event={"ID":"215dfd5c-6d06-42ef-89c7-6d40910fbf28","Type":"ContainerStarted","Data":"e0b6fa7f3f62deb21226eab9359b6cbd960d2e58ccd84ff89b6776767266058e"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.589914 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podStartSLOduration=29.051573909 podStartE2EDuration="29.589887304s" podCreationTimestamp="2026-03-09 18:47:12 +0000 UTC" firstStartedPulling="2026-03-09 18:47:38.179733526 +0000 UTC m=+1339.522205924" lastFinishedPulling="2026-03-09 18:47:38.718046921 +0000 UTC m=+1340.060519319" observedRunningTime="2026-03-09 18:47:41.57024533 +0000 UTC m=+1342.912717728" watchObservedRunningTime="2026-03-09 18:47:41.589887304 +0000 UTC m=+1342.932359702" Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.605225 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d177b683-bf5e-4048-a0cf-2b7e6176f2ed","Type":"ContainerStarted","Data":"6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670"} Mar 09 18:47:41 crc kubenswrapper[4750]: I0309 18:47:41.625287 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-59wvm" podStartSLOduration=6.466421885 podStartE2EDuration="38.625256664s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="2026-03-09 18:47:06.56021356 +0000 UTC m=+1307.902685968" lastFinishedPulling="2026-03-09 18:47:38.719048349 +0000 UTC m=+1340.061520747" observedRunningTime="2026-03-09 18:47:41.598778045 +0000 UTC m=+1342.941250453" watchObservedRunningTime="2026-03-09 18:47:41.625256664 +0000 UTC m=+1342.967729062" Mar 09 18:47:42 crc kubenswrapper[4750]: I0309 18:47:42.646663 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748768969-j7m74" event={"ID":"72dade1d-7fb7-4118-a665-7b2fdfc56a9e","Type":"ContainerStarted","Data":"89b0977a0ec3c9f5642c50c7e3f122d603979f0c03ed7c79afc7d8bf99ce5b5b"} Mar 09 18:47:42 crc kubenswrapper[4750]: I0309 18:47:42.671206 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" event={"ID":"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e","Type":"ContainerStarted","Data":"c3d6fa510aa940a1bd25fc30b24a26fc38dcbd0e75626c836c87f55051ab4d09"} Mar 09 18:47:42 crc kubenswrapper[4750]: I0309 18:47:42.677007 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e","Type":"ContainerStarted","Data":"f4f3af90d447b2735d6453a83cbcee8ac25b9534a12865d87df3bc2bde9048a6"} Mar 09 18:47:42 crc kubenswrapper[4750]: I0309 18:47:42.692083 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d177b683-bf5e-4048-a0cf-2b7e6176f2ed","Type":"ContainerStarted","Data":"976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1"} Mar 09 18:47:42 crc kubenswrapper[4750]: I0309 18:47:42.724174 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=17.72414707 podStartE2EDuration="17.72414707s" podCreationTimestamp="2026-03-09 18:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:42.714044225 +0000 UTC m=+1344.056516623" watchObservedRunningTime="2026-03-09 18:47:42.72414707 +0000 UTC m=+1344.066619468" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.348238 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.348608 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.437624 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.437732 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.724926 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e","Type":"ContainerStarted","Data":"2a713089dd3eef866ef549e6aae03e6e727a581f21b374c170ba07ed6358385b"} Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.740930 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748768969-j7m74" event={"ID":"72dade1d-7fb7-4118-a665-7b2fdfc56a9e","Type":"ContainerStarted","Data":"0d8051df72ad676824e604b83bb7b43891baff0d89ef21f42ae3b5cb0c1ab46b"} Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.742137 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-748768969-j7m74" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.742166 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.777312 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=18.777284831 podStartE2EDuration="18.777284831s" podCreationTimestamp="2026-03-09 18:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:43.761066791 +0000 UTC m=+1345.103539189" watchObservedRunningTime="2026-03-09 18:47:43.777284831 +0000 UTC m=+1345.119757229" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.798516 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" podStartSLOduration=6.798490217 podStartE2EDuration="6.798490217s" podCreationTimestamp="2026-03-09 18:47:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:43.788177447 +0000 UTC m=+1345.130649855" watchObservedRunningTime="2026-03-09 18:47:43.798490217 +0000 UTC m=+1345.140962615" Mar 09 18:47:43 crc kubenswrapper[4750]: I0309 18:47:43.815691 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-748768969-j7m74" podStartSLOduration=4.815604053 podStartE2EDuration="4.815604053s" podCreationTimestamp="2026-03-09 18:47:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:43.811306236 +0000 UTC m=+1345.153778654" watchObservedRunningTime="2026-03-09 18:47:43.815604053 +0000 UTC m=+1345.158076451" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.061809 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.118279 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.571300 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.611345 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.752603 4750 generic.go:334] "Generic (PLEG): container finished" podID="2afe8078-9724-4ab3-b5ee-79c978ce9bea" containerID="8e3d7cd0c864c75c3c552a0a15ab5245d401dcf5c5c8c397a62a26b48376969d" exitCode=0 Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.752696 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsgm7" event={"ID":"2afe8078-9724-4ab3-b5ee-79c978ce9bea","Type":"ContainerDied","Data":"8e3d7cd0c864c75c3c552a0a15ab5245d401dcf5c5c8c397a62a26b48376969d"} Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.753891 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.788733 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.795930 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.825704 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:44 crc kubenswrapper[4750]: I0309 18:47:44.843938 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:45 crc kubenswrapper[4750]: I0309 18:47:45.764627 4750 generic.go:334] "Generic (PLEG): container finished" podID="250d3d1c-a665-4e77-8ac0-15768f6a25f1" containerID="7ccb2dde1714b67016d32aa2bf42313d9a79e120ae7ad1e8ed5ac4e14697f4ea" exitCode=0 Mar 09 18:47:45 crc kubenswrapper[4750]: I0309 18:47:45.764837 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t8lkt" event={"ID":"250d3d1c-a665-4e77-8ac0-15768f6a25f1","Type":"ContainerDied","Data":"7ccb2dde1714b67016d32aa2bf42313d9a79e120ae7ad1e8ed5ac4e14697f4ea"} Mar 09 18:47:45 crc kubenswrapper[4750]: I0309 18:47:45.771192 4750 generic.go:334] "Generic (PLEG): container finished" podID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerID="793969d3d3e5cab9242146e23eb4c826c91cafe7e607d142e01d40eb07324c7c" exitCode=1 Mar 09 18:47:45 crc kubenswrapper[4750]: I0309 18:47:45.772768 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"9d32a26f-bd14-41a7-872b-e102ea95f696","Type":"ContainerDied","Data":"793969d3d3e5cab9242146e23eb4c826c91cafe7e607d142e01d40eb07324c7c"} Mar 09 18:47:45 crc kubenswrapper[4750]: I0309 18:47:45.772950 4750 scope.go:117] "RemoveContainer" containerID="793969d3d3e5cab9242146e23eb4c826c91cafe7e607d142e01d40eb07324c7c" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.243343 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.243591 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.289350 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.289644 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.518528 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.518981 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.570328 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.588865 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.788724 4750 generic.go:334] "Generic (PLEG): container finished" podID="215dfd5c-6d06-42ef-89c7-6d40910fbf28" containerID="e0b6fa7f3f62deb21226eab9359b6cbd960d2e58ccd84ff89b6776767266058e" exitCode=0 Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.790512 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-59wvm" event={"ID":"215dfd5c-6d06-42ef-89c7-6d40910fbf28","Type":"ContainerDied","Data":"e0b6fa7f3f62deb21226eab9359b6cbd960d2e58ccd84ff89b6776767266058e"} Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.790598 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.791514 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="53dae358-d7dd-4823-8369-d548e214333e" containerName="watcher-applier" containerID="cri-o://403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d" gracePeriod=30 Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.791762 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.791833 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:46 crc kubenswrapper[4750]: I0309 18:47:46.791849 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 09 18:47:47 crc kubenswrapper[4750]: I0309 18:47:47.650888 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:47:47 crc kubenswrapper[4750]: I0309 18:47:47.732526 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5cd9697c-zjwc7"] Mar 09 18:47:47 crc kubenswrapper[4750]: I0309 18:47:47.732820 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" podUID="031456d8-c16a-49fa-9b19-b22060292ac3" containerName="dnsmasq-dns" containerID="cri-o://4532b42f7be8a187289ed33dac72c539a14b331df2223ee86ccbbd600a95d28d" gracePeriod=10 Mar 09 18:47:48 crc kubenswrapper[4750]: I0309 18:47:48.819896 4750 generic.go:334] "Generic (PLEG): container finished" podID="031456d8-c16a-49fa-9b19-b22060292ac3" containerID="4532b42f7be8a187289ed33dac72c539a14b331df2223ee86ccbbd600a95d28d" exitCode=0 Mar 09 18:47:48 crc kubenswrapper[4750]: I0309 18:47:48.819981 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" event={"ID":"031456d8-c16a-49fa-9b19-b22060292ac3","Type":"ContainerDied","Data":"4532b42f7be8a187289ed33dac72c539a14b331df2223ee86ccbbd600a95d28d"} Mar 09 18:47:48 crc kubenswrapper[4750]: I0309 18:47:48.823251 4750 generic.go:334] "Generic (PLEG): container finished" podID="53dae358-d7dd-4823-8369-d548e214333e" containerID="403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d" exitCode=0 Mar 09 18:47:48 crc kubenswrapper[4750]: I0309 18:47:48.823348 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:47:48 crc kubenswrapper[4750]: I0309 18:47:48.823566 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"53dae358-d7dd-4823-8369-d548e214333e","Type":"ContainerDied","Data":"403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d"} Mar 09 18:47:48 crc kubenswrapper[4750]: I0309 18:47:48.823707 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.499573 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.516512 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.557620 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:49 crc kubenswrapper[4750]: E0309 18:47:49.571919 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d is running failed: container process not found" containerID="403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 09 18:47:49 crc kubenswrapper[4750]: E0309 18:47:49.576035 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d is running failed: container process not found" containerID="403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 09 18:47:49 crc kubenswrapper[4750]: E0309 18:47:49.579581 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d is running failed: container process not found" containerID="403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 09 18:47:49 crc kubenswrapper[4750]: E0309 18:47:49.579687 4750 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d is running failed: container process not found" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="53dae358-d7dd-4823-8369-d548e214333e" containerName="watcher-applier" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.655841 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-combined-ca-bundle\") pod \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.656251 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz94p\" (UniqueName: \"kubernetes.io/projected/2afe8078-9724-4ab3-b5ee-79c978ce9bea-kube-api-access-rz94p\") pod \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.656746 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2afe8078-9724-4ab3-b5ee-79c978ce9bea-logs\") pod \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.656782 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-config-data\") pod \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.656849 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-combined-ca-bundle\") pod \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.656907 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-db-sync-config-data\") pod \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.656944 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-scripts\") pod \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\" (UID: \"2afe8078-9724-4ab3-b5ee-79c978ce9bea\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.656978 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-credential-keys\") pod \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.657020 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbxrg\" (UniqueName: \"kubernetes.io/projected/215dfd5c-6d06-42ef-89c7-6d40910fbf28-kube-api-access-hbxrg\") pod \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\" (UID: \"215dfd5c-6d06-42ef-89c7-6d40910fbf28\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.657048 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-fernet-keys\") pod \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.657073 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-combined-ca-bundle\") pod \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.657101 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-scripts\") pod \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.657121 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfwf6\" (UniqueName: \"kubernetes.io/projected/250d3d1c-a665-4e77-8ac0-15768f6a25f1-kube-api-access-tfwf6\") pod \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.657152 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-config-data\") pod \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\" (UID: \"250d3d1c-a665-4e77-8ac0-15768f6a25f1\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.665811 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-scripts" (OuterVolumeSpecName: "scripts") pod "2afe8078-9724-4ab3-b5ee-79c978ce9bea" (UID: "2afe8078-9724-4ab3-b5ee-79c978ce9bea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.666850 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2afe8078-9724-4ab3-b5ee-79c978ce9bea-logs" (OuterVolumeSpecName: "logs") pod "2afe8078-9724-4ab3-b5ee-79c978ce9bea" (UID: "2afe8078-9724-4ab3-b5ee-79c978ce9bea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.679889 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "215dfd5c-6d06-42ef-89c7-6d40910fbf28" (UID: "215dfd5c-6d06-42ef-89c7-6d40910fbf28"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.684748 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "250d3d1c-a665-4e77-8ac0-15768f6a25f1" (UID: "250d3d1c-a665-4e77-8ac0-15768f6a25f1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.695982 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-scripts" (OuterVolumeSpecName: "scripts") pod "250d3d1c-a665-4e77-8ac0-15768f6a25f1" (UID: "250d3d1c-a665-4e77-8ac0-15768f6a25f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.701825 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "250d3d1c-a665-4e77-8ac0-15768f6a25f1" (UID: "250d3d1c-a665-4e77-8ac0-15768f6a25f1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.777014 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2afe8078-9724-4ab3-b5ee-79c978ce9bea-kube-api-access-rz94p" (OuterVolumeSpecName: "kube-api-access-rz94p") pod "2afe8078-9724-4ab3-b5ee-79c978ce9bea" (UID: "2afe8078-9724-4ab3-b5ee-79c978ce9bea"). InnerVolumeSpecName "kube-api-access-rz94p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.777569 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/215dfd5c-6d06-42ef-89c7-6d40910fbf28-kube-api-access-hbxrg" (OuterVolumeSpecName: "kube-api-access-hbxrg") pod "215dfd5c-6d06-42ef-89c7-6d40910fbf28" (UID: "215dfd5c-6d06-42ef-89c7-6d40910fbf28"). InnerVolumeSpecName "kube-api-access-hbxrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783061 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz94p\" (UniqueName: \"kubernetes.io/projected/2afe8078-9724-4ab3-b5ee-79c978ce9bea-kube-api-access-rz94p\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783108 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2afe8078-9724-4ab3-b5ee-79c978ce9bea-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783125 4750 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783138 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783150 4750 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783164 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbxrg\" (UniqueName: \"kubernetes.io/projected/215dfd5c-6d06-42ef-89c7-6d40910fbf28-kube-api-access-hbxrg\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783176 4750 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.783188 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.795426 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/250d3d1c-a665-4e77-8ac0-15768f6a25f1-kube-api-access-tfwf6" (OuterVolumeSpecName: "kube-api-access-tfwf6") pod "250d3d1c-a665-4e77-8ac0-15768f6a25f1" (UID: "250d3d1c-a665-4e77-8ac0-15768f6a25f1"). InnerVolumeSpecName "kube-api-access-tfwf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.801832 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "215dfd5c-6d06-42ef-89c7-6d40910fbf28" (UID: "215dfd5c-6d06-42ef-89c7-6d40910fbf28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.803616 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-config-data" (OuterVolumeSpecName: "config-data") pod "250d3d1c-a665-4e77-8ac0-15768f6a25f1" (UID: "250d3d1c-a665-4e77-8ac0-15768f6a25f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.806350 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-config-data" (OuterVolumeSpecName: "config-data") pod "2afe8078-9724-4ab3-b5ee-79c978ce9bea" (UID: "2afe8078-9724-4ab3-b5ee-79c978ce9bea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.825722 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2afe8078-9724-4ab3-b5ee-79c978ce9bea" (UID: "2afe8078-9724-4ab3-b5ee-79c978ce9bea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.859086 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.884305 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.884335 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfwf6\" (UniqueName: \"kubernetes.io/projected/250d3d1c-a665-4e77-8ac0-15768f6a25f1-kube-api-access-tfwf6\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.884346 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.884355 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/215dfd5c-6d06-42ef-89c7-6d40910fbf28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.884364 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afe8078-9724-4ab3-b5ee-79c978ce9bea-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.887285 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "250d3d1c-a665-4e77-8ac0-15768f6a25f1" (UID: "250d3d1c-a665-4e77-8ac0-15768f6a25f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.889996 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t8lkt" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.890786 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t8lkt" event={"ID":"250d3d1c-a665-4e77-8ac0-15768f6a25f1","Type":"ContainerDied","Data":"2bfc3ac1b91b7658d08225b83ff4b7111fc8a4daaa14e54bc8ec8944f062338a"} Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.890839 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bfc3ac1b91b7658d08225b83ff4b7111fc8a4daaa14e54bc8ec8944f062338a" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.934187 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" event={"ID":"031456d8-c16a-49fa-9b19-b22060292ac3","Type":"ContainerDied","Data":"f531f7fe33696cfd23d1d0cf3fbef422e8892bd1ff13db52684a150209325314"} Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.934257 4750 scope.go:117] "RemoveContainer" containerID="4532b42f7be8a187289ed33dac72c539a14b331df2223ee86ccbbd600a95d28d" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.934435 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5cd9697c-zjwc7" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.941944 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsgm7" event={"ID":"2afe8078-9724-4ab3-b5ee-79c978ce9bea","Type":"ContainerDied","Data":"d9ddcbfb1c220ea4f30459640ea4b9f2c20d0557c721a088be9aa53cbc747177"} Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.941994 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9ddcbfb1c220ea4f30459640ea4b9f2c20d0557c721a088be9aa53cbc747177" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.942144 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsgm7" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.944587 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-59wvm" event={"ID":"215dfd5c-6d06-42ef-89c7-6d40910fbf28","Type":"ContainerDied","Data":"5308a9c506d492aaa4a052237cfde65a23247123a3f9d9159f5aa133183ca01b"} Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.944651 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5308a9c506d492aaa4a052237cfde65a23247123a3f9d9159f5aa133183ca01b" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.944858 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-59wvm" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.963107 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.991108 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-nb\") pod \"031456d8-c16a-49fa-9b19-b22060292ac3\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.991662 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnnbn\" (UniqueName: \"kubernetes.io/projected/031456d8-c16a-49fa-9b19-b22060292ac3-kube-api-access-rnnbn\") pod \"031456d8-c16a-49fa-9b19-b22060292ac3\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.991766 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-swift-storage-0\") pod \"031456d8-c16a-49fa-9b19-b22060292ac3\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.991874 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-svc\") pod \"031456d8-c16a-49fa-9b19-b22060292ac3\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.992065 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-config\") pod \"031456d8-c16a-49fa-9b19-b22060292ac3\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.992309 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-sb\") pod \"031456d8-c16a-49fa-9b19-b22060292ac3\" (UID: \"031456d8-c16a-49fa-9b19-b22060292ac3\") " Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.993567 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/250d3d1c-a665-4e77-8ac0-15768f6a25f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:49 crc kubenswrapper[4750]: I0309 18:47:49.999334 4750 scope.go:117] "RemoveContainer" containerID="53113e128ce42ad1d1f8b5c4174cd3215f52c58caa7e388816634f12bc81f2db" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.009709 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/031456d8-c16a-49fa-9b19-b22060292ac3-kube-api-access-rnnbn" (OuterVolumeSpecName: "kube-api-access-rnnbn") pod "031456d8-c16a-49fa-9b19-b22060292ac3" (UID: "031456d8-c16a-49fa-9b19-b22060292ac3"). InnerVolumeSpecName "kube-api-access-rnnbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.095302 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53dae358-d7dd-4823-8369-d548e214333e-logs\") pod \"53dae358-d7dd-4823-8369-d548e214333e\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.095675 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k6kr\" (UniqueName: \"kubernetes.io/projected/53dae358-d7dd-4823-8369-d548e214333e-kube-api-access-2k6kr\") pod \"53dae358-d7dd-4823-8369-d548e214333e\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.097204 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-combined-ca-bundle\") pod \"53dae358-d7dd-4823-8369-d548e214333e\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.097281 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-config-data\") pod \"53dae358-d7dd-4823-8369-d548e214333e\" (UID: \"53dae358-d7dd-4823-8369-d548e214333e\") " Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.095880 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53dae358-d7dd-4823-8369-d548e214333e-logs" (OuterVolumeSpecName: "logs") pod "53dae358-d7dd-4823-8369-d548e214333e" (UID: "53dae358-d7dd-4823-8369-d548e214333e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.098149 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53dae358-d7dd-4823-8369-d548e214333e-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.098915 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnnbn\" (UniqueName: \"kubernetes.io/projected/031456d8-c16a-49fa-9b19-b22060292ac3-kube-api-access-rnnbn\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.115099 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53dae358-d7dd-4823-8369-d548e214333e-kube-api-access-2k6kr" (OuterVolumeSpecName: "kube-api-access-2k6kr") pod "53dae358-d7dd-4823-8369-d548e214333e" (UID: "53dae358-d7dd-4823-8369-d548e214333e"). InnerVolumeSpecName "kube-api-access-2k6kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.169195 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "031456d8-c16a-49fa-9b19-b22060292ac3" (UID: "031456d8-c16a-49fa-9b19-b22060292ac3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.172026 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-config" (OuterVolumeSpecName: "config") pod "031456d8-c16a-49fa-9b19-b22060292ac3" (UID: "031456d8-c16a-49fa-9b19-b22060292ac3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.204999 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.205025 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k6kr\" (UniqueName: \"kubernetes.io/projected/53dae358-d7dd-4823-8369-d548e214333e-kube-api-access-2k6kr\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.205035 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.205985 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "031456d8-c16a-49fa-9b19-b22060292ac3" (UID: "031456d8-c16a-49fa-9b19-b22060292ac3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.211824 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53dae358-d7dd-4823-8369-d548e214333e" (UID: "53dae358-d7dd-4823-8369-d548e214333e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.216842 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-config-data" (OuterVolumeSpecName: "config-data") pod "53dae358-d7dd-4823-8369-d548e214333e" (UID: "53dae358-d7dd-4823-8369-d548e214333e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.229585 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "031456d8-c16a-49fa-9b19-b22060292ac3" (UID: "031456d8-c16a-49fa-9b19-b22060292ac3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.254364 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "031456d8-c16a-49fa-9b19-b22060292ac3" (UID: "031456d8-c16a-49fa-9b19-b22060292ac3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.310462 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.310517 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53dae358-d7dd-4823-8369-d548e214333e-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.310528 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.310540 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.310551 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031456d8-c16a-49fa-9b19-b22060292ac3-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.592223 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5cd9697c-zjwc7"] Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.612993 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f5cd9697c-zjwc7"] Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.656395 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6594b67854-xmvs4"] Mar 09 18:47:50 crc kubenswrapper[4750]: E0309 18:47:50.656856 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="215dfd5c-6d06-42ef-89c7-6d40910fbf28" containerName="barbican-db-sync" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.656873 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="215dfd5c-6d06-42ef-89c7-6d40910fbf28" containerName="barbican-db-sync" Mar 09 18:47:50 crc kubenswrapper[4750]: E0309 18:47:50.656892 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250d3d1c-a665-4e77-8ac0-15768f6a25f1" containerName="keystone-bootstrap" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.656899 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="250d3d1c-a665-4e77-8ac0-15768f6a25f1" containerName="keystone-bootstrap" Mar 09 18:47:50 crc kubenswrapper[4750]: E0309 18:47:50.656915 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="031456d8-c16a-49fa-9b19-b22060292ac3" containerName="dnsmasq-dns" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.656923 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="031456d8-c16a-49fa-9b19-b22060292ac3" containerName="dnsmasq-dns" Mar 09 18:47:50 crc kubenswrapper[4750]: E0309 18:47:50.656939 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="031456d8-c16a-49fa-9b19-b22060292ac3" containerName="init" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.656945 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="031456d8-c16a-49fa-9b19-b22060292ac3" containerName="init" Mar 09 18:47:50 crc kubenswrapper[4750]: E0309 18:47:50.656960 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53dae358-d7dd-4823-8369-d548e214333e" containerName="watcher-applier" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.656966 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="53dae358-d7dd-4823-8369-d548e214333e" containerName="watcher-applier" Mar 09 18:47:50 crc kubenswrapper[4750]: E0309 18:47:50.656980 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2afe8078-9724-4ab3-b5ee-79c978ce9bea" containerName="placement-db-sync" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.656985 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2afe8078-9724-4ab3-b5ee-79c978ce9bea" containerName="placement-db-sync" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.657163 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="250d3d1c-a665-4e77-8ac0-15768f6a25f1" containerName="keystone-bootstrap" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.657179 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2afe8078-9724-4ab3-b5ee-79c978ce9bea" containerName="placement-db-sync" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.657189 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="53dae358-d7dd-4823-8369-d548e214333e" containerName="watcher-applier" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.657201 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="215dfd5c-6d06-42ef-89c7-6d40910fbf28" containerName="barbican-db-sync" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.657213 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="031456d8-c16a-49fa-9b19-b22060292ac3" containerName="dnsmasq-dns" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.657860 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.662207 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9flmk" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.663063 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.663210 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.663314 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.663344 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.668890 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.668989 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6594b67854-xmvs4"] Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.779539 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-84fc97847d-hvzvz"] Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.781718 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.792819 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.793027 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.793228 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xzm7g" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.793342 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.793545 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.800855 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84fc97847d-hvzvz"] Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825187 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwd59\" (UniqueName: \"kubernetes.io/projected/f1171881-c394-4042-b092-a6935ed182df-kube-api-access-xwd59\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825256 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-combined-ca-bundle\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825286 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-fernet-keys\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825303 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-scripts\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825343 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-config-data\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825385 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-public-tls-certs\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825417 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-internal-tls-certs\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.825446 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-credential-keys\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.918523 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-fb5b94469-qjhvw"] Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.920608 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929122 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-public-tls-certs\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929173 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-scripts\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929213 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-combined-ca-bundle\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929230 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-internal-tls-certs\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929256 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-credential-keys\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929274 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-internal-tls-certs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929297 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwd59\" (UniqueName: \"kubernetes.io/projected/f1171881-c394-4042-b092-a6935ed182df-kube-api-access-xwd59\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-combined-ca-bundle\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929368 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-config-data\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929385 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-fernet-keys\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929402 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-scripts\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929424 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-public-tls-certs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929448 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlbqn\" (UniqueName: \"kubernetes.io/projected/edd199bd-36e4-4f32-a4cf-237d7a88011a-kube-api-access-vlbqn\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929478 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-config-data\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.929496 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd199bd-36e4-4f32-a4cf-237d7a88011a-logs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.937531 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-public-tls-certs\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.939831 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-combined-ca-bundle\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.944374 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-internal-tls-certs\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.953732 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-66cf65f968-kqpdg"] Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.955389 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.971814 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.971225 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.972436 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pq62m" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.977553 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-fernet-keys\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.978343 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-config-data\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.979174 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 09 18:47:50 crc kubenswrapper[4750]: I0309 18:47:50.980017 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fb5b94469-qjhvw"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:50.995608 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-credential-keys\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.016202 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwd59\" (UniqueName: \"kubernetes.io/projected/f1171881-c394-4042-b092-a6935ed182df-kube-api-access-xwd59\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.016746 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66cf65f968-kqpdg"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.028430 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1171881-c394-4042-b092-a6935ed182df-scripts\") pod \"keystone-6594b67854-xmvs4\" (UID: \"f1171881-c394-4042-b092-a6935ed182df\") " pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.030844 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-scripts\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.030912 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-combined-ca-bundle\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.030946 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data-custom\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.030964 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-combined-ca-bundle\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.030996 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.031021 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-internal-tls-certs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.031077 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840a6d14-b199-41f2-8adf-124089a1ec27-logs\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.031101 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-config-data\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.031121 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-public-tls-certs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.031148 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlbqn\" (UniqueName: \"kubernetes.io/projected/edd199bd-36e4-4f32-a4cf-237d7a88011a-kube-api-access-vlbqn\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.031179 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd199bd-36e4-4f32-a4cf-237d7a88011a-logs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.031211 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwrr9\" (UniqueName: \"kubernetes.io/projected/840a6d14-b199-41f2-8adf-124089a1ec27-kube-api-access-dwrr9\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.051923 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd199bd-36e4-4f32-a4cf-237d7a88011a-logs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.058606 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-config-data\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.065797 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-scripts\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.069764 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fc5774769-shw5p"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.078018 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.084813 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-internal-tls-certs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.085776 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fc5774769-shw5p"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.089810 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-combined-ca-bundle\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.091148 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-public-tls-certs\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.091716 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlbqn\" (UniqueName: \"kubernetes.io/projected/edd199bd-36e4-4f32-a4cf-237d7a88011a-kube-api-access-vlbqn\") pod \"placement-84fc97847d-hvzvz\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.172125 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.173645 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-89k9z" event={"ID":"fec39559-ddae-4e19-8941-0fc3209084cd","Type":"ContainerStarted","Data":"fa00c999ec13985a6d12a0ee552edbc27058d1f0a0de89d2e82c9c53574c40f7"} Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.205074 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data-custom\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.215596 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwrr9\" (UniqueName: \"kubernetes.io/projected/840a6d14-b199-41f2-8adf-124089a1ec27-kube-api-access-dwrr9\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.233617 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerStarted","Data":"5e6269255415c111f968e23ad15921e57205996484e746f79d35fde5ab0b990e"} Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.233664 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-combined-ca-bundle\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.233883 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data-custom\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.233991 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.234070 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90fb1eef-5754-4823-8fbc-693ebeb8deb9-logs\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.234167 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg5zq\" (UniqueName: \"kubernetes.io/projected/90fb1eef-5754-4823-8fbc-693ebeb8deb9-kube-api-access-jg5zq\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.234267 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.234436 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840a6d14-b199-41f2-8adf-124089a1ec27-logs\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.234595 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-combined-ca-bundle\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.235624 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840a6d14-b199-41f2-8adf-124089a1ec27-logs\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.243730 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"53dae358-d7dd-4823-8369-d548e214333e","Type":"ContainerDied","Data":"a9b4eddf4191e6e62d4b90b10776648af8ca32f0e90a7242f84ef0704cf64949"} Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.243830 4750 scope.go:117] "RemoveContainer" containerID="403726f8d1927333ba3bf99678c04cfda8a5df9b8aaed197c313b26bbf53af6d" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.244140 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.257386 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-combined-ca-bundle\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.262057 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwrr9\" (UniqueName: \"kubernetes.io/projected/840a6d14-b199-41f2-8adf-124089a1ec27-kube-api-access-dwrr9\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.264486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data-custom\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.267190 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data\") pod \"barbican-worker-fb5b94469-qjhvw\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.277732 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"9d32a26f-bd14-41a7-872b-e102ea95f696","Type":"ContainerStarted","Data":"0c4186de4c7462289849baaa47e27468f2c826e928ce6d9991a34baefed76f23"} Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.277951 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerName="watcher-decision-engine" containerID="cri-o://0c4186de4c7462289849baaa47e27468f2c826e928ce6d9991a34baefed76f23" gracePeriod=30 Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.288001 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.351520 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data-custom\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.351609 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-swift-storage-0\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.351707 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snxst\" (UniqueName: \"kubernetes.io/projected/5be93415-f615-42a7-948f-68829841543e-kube-api-access-snxst\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.351816 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-sb\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.351919 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90fb1eef-5754-4823-8fbc-693ebeb8deb9-logs\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.351998 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg5zq\" (UniqueName: \"kubernetes.io/projected/90fb1eef-5754-4823-8fbc-693ebeb8deb9-kube-api-access-jg5zq\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.352024 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-nb\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.352065 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.353189 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90fb1eef-5754-4823-8fbc-693ebeb8deb9-logs\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.354144 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-svc\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.354482 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-combined-ca-bundle\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.354595 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-config\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.364161 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data-custom\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.371845 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.376693 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-combined-ca-bundle\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.418760 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg5zq\" (UniqueName: \"kubernetes.io/projected/90fb1eef-5754-4823-8fbc-693ebeb8deb9-kube-api-access-jg5zq\") pod \"barbican-keystone-listener-66cf65f968-kqpdg\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.459688 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-sb\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.460103 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-nb\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.460274 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-svc\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.463827 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-config\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.464086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-swift-storage-0\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.464290 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snxst\" (UniqueName: \"kubernetes.io/projected/5be93415-f615-42a7-948f-68829841543e-kube-api-access-snxst\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.466398 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-nb\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.467075 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-swift-storage-0\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.467951 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-svc\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.468983 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-config\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.484273 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-sb\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.506446 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="031456d8-c16a-49fa-9b19-b22060292ac3" path="/var/lib/kubelet/pods/031456d8-c16a-49fa-9b19-b22060292ac3/volumes" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.507350 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-577799fd8d-66jqt"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.517452 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-89k9z" podStartSLOduration=4.521788264 podStartE2EDuration="48.517420887s" podCreationTimestamp="2026-03-09 18:47:03 +0000 UTC" firstStartedPulling="2026-03-09 18:47:05.560206532 +0000 UTC m=+1306.902678930" lastFinishedPulling="2026-03-09 18:47:49.555839155 +0000 UTC m=+1350.898311553" observedRunningTime="2026-03-09 18:47:51.224088158 +0000 UTC m=+1352.566560556" watchObservedRunningTime="2026-03-09 18:47:51.517420887 +0000 UTC m=+1352.859893285" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.525054 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-577799fd8d-66jqt"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.525400 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.528155 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.541011 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snxst\" (UniqueName: \"kubernetes.io/projected/5be93415-f615-42a7-948f-68829841543e-kube-api-access-snxst\") pod \"dnsmasq-dns-fc5774769-shw5p\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.560835 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.561380 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-54d9b479b6-f2tns"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.563564 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.565816 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data-custom\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.565903 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zv8r\" (UniqueName: \"kubernetes.io/projected/cc0b5242-afa5-4413-9887-432846908b1d-kube-api-access-7zv8r\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.565938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-combined-ca-bundle\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.565983 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0b5242-afa5-4413-9887-432846908b1d-logs\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.566008 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.604743 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.636557 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54d9b479b6-f2tns"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.646637 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.668000 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data-custom\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.668082 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zv8r\" (UniqueName: \"kubernetes.io/projected/cc0b5242-afa5-4413-9887-432846908b1d-kube-api-access-7zv8r\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.668114 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-combined-ca-bundle\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.668155 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0b5242-afa5-4413-9887-432846908b1d-logs\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.668179 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.676323 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0b5242-afa5-4413-9887-432846908b1d-logs\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.682724 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data-custom\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.684896 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.688356 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-combined-ca-bundle\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.736243 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zv8r\" (UniqueName: \"kubernetes.io/projected/cc0b5242-afa5-4413-9887-432846908b1d-kube-api-access-7zv8r\") pod \"barbican-api-577799fd8d-66jqt\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.736331 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76cfd56bfc-9l7jh"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.741309 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.771862 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jffwc\" (UniqueName: \"kubernetes.io/projected/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-kube-api-access-jffwc\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.771896 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-combined-ca-bundle\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.771957 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-config-data-custom\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.771981 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-config-data\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.772021 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-logs\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.793035 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76cfd56bfc-9l7jh"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.829440 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.849122 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.876771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-combined-ca-bundle\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.876862 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vnlv\" (UniqueName: \"kubernetes.io/projected/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-kube-api-access-9vnlv\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.876897 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-logs\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.876940 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-config-data-custom\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.876966 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-combined-ca-bundle\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.876998 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-config-data\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.877025 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-config-data-custom\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.877074 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-logs\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.877102 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-config-data\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.877166 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jffwc\" (UniqueName: \"kubernetes.io/projected/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-kube-api-access-jffwc\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.882128 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-combined-ca-bundle\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.889201 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-logs\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.889964 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.890815 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-config-data\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.893579 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-config-data-custom\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.917345 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.927373 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.928320 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jffwc\" (UniqueName: \"kubernetes.io/projected/c44181c7-bb39-45a8-ad11-11f4aab9f8b7-kube-api-access-jffwc\") pod \"barbican-keystone-listener-54d9b479b6-f2tns\" (UID: \"c44181c7-bb39-45a8-ad11-11f4aab9f8b7\") " pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.939953 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.979950 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vnlv\" (UniqueName: \"kubernetes.io/projected/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-kube-api-access-9vnlv\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.980012 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-logs\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.980080 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-combined-ca-bundle\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.980123 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-config-data-custom\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.980185 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-config-data\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.990700 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-79b7cc4cf4-w25dt"] Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.991520 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-logs\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:51 crc kubenswrapper[4750]: I0309 18:47:51.992655 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:51.998980 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-combined-ca-bundle\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.001333 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-config-data\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.003358 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-config-data-custom\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.028919 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.042519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vnlv\" (UniqueName: \"kubernetes.io/projected/968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0-kube-api-access-9vnlv\") pod \"barbican-worker-76cfd56bfc-9l7jh\" (UID: \"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0\") " pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.055303 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b7cc4cf4-w25dt"] Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096291 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/150ab1b1-103a-463a-9c5d-fc4231e42674-logs\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096355 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9mwh\" (UniqueName: \"kubernetes.io/projected/c857b2a0-c8c1-452f-99af-a6affa05c8c1-kube-api-access-m9mwh\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096380 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096461 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-config-data\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096489 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096509 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data-custom\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096528 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v2xw\" (UniqueName: \"kubernetes.io/projected/150ab1b1-103a-463a-9c5d-fc4231e42674-kube-api-access-4v2xw\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096567 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c857b2a0-c8c1-452f-99af-a6affa05c8c1-logs\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.096599 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-combined-ca-bundle\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.123480 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.151008 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76cfd56bfc-9l7jh" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.199240 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-config-data\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.199299 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.199321 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data-custom\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.199340 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v2xw\" (UniqueName: \"kubernetes.io/projected/150ab1b1-103a-463a-9c5d-fc4231e42674-kube-api-access-4v2xw\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.202385 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c857b2a0-c8c1-452f-99af-a6affa05c8c1-logs\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.202423 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-combined-ca-bundle\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.205365 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c857b2a0-c8c1-452f-99af-a6affa05c8c1-logs\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.205754 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/150ab1b1-103a-463a-9c5d-fc4231e42674-logs\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.205806 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9mwh\" (UniqueName: \"kubernetes.io/projected/c857b2a0-c8c1-452f-99af-a6affa05c8c1-kube-api-access-m9mwh\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.205831 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.206545 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84fc97847d-hvzvz"] Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.207512 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/150ab1b1-103a-463a-9c5d-fc4231e42674-logs\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.212342 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.214882 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-combined-ca-bundle\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.219277 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-config-data\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.224546 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.231848 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data-custom\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.238160 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v2xw\" (UniqueName: \"kubernetes.io/projected/150ab1b1-103a-463a-9c5d-fc4231e42674-kube-api-access-4v2xw\") pod \"barbican-api-79b7cc4cf4-w25dt\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.242298 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9mwh\" (UniqueName: \"kubernetes.io/projected/c857b2a0-c8c1-452f-99af-a6affa05c8c1-kube-api-access-m9mwh\") pod \"watcher-applier-0\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.314453 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.323057 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84fc97847d-hvzvz" event={"ID":"edd199bd-36e4-4f32-a4cf-237d7a88011a","Type":"ContainerStarted","Data":"dd07a98c706b499332baba6944b47826b2b92713f43d2ade90b9ba0a549bcd34"} Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.374303 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.520747 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fb5b94469-qjhvw"] Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.543603 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6594b67854-xmvs4"] Mar 09 18:47:52 crc kubenswrapper[4750]: I0309 18:47:52.732339 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66cf65f968-kqpdg"] Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.043350 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.043840 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.049768 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.052894 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.053011 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.149234 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-577799fd8d-66jqt"] Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.172799 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fc5774769-shw5p"] Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.411353 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.167:8443: connect: connection refused" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.419957 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53dae358-d7dd-4823-8369-d548e214333e" path="/var/lib/kubelet/pods/53dae358-d7dd-4823-8369-d548e214333e/volumes" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.421697 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76cfd56bfc-9l7jh"] Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.421729 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6594b67854-xmvs4" event={"ID":"f1171881-c394-4042-b092-a6935ed182df","Type":"ContainerStarted","Data":"3f04ab411538d2c2e8fc03156b28f583174d318af27c0a27e114fe9770a0a82d"} Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.421748 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-577799fd8d-66jqt" event={"ID":"cc0b5242-afa5-4413-9887-432846908b1d","Type":"ContainerStarted","Data":"c131d8c28d81b78059647080be8d3146e50b9bbde8c4bdcb43e225ca53e638bb"} Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.421760 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb5b94469-qjhvw" event={"ID":"840a6d14-b199-41f2-8adf-124089a1ec27","Type":"ContainerStarted","Data":"4713f798673c0dc6bdcb80a7894a711720f73c75a4c12e69567d2c83767932d1"} Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.425619 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54d9b479b6-f2tns"] Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.429782 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fc5774769-shw5p" event={"ID":"5be93415-f615-42a7-948f-68829841543e","Type":"ContainerStarted","Data":"794ab5ae835235f7e7658e998c2546d0f8d0c033ee19c70e79f6e72aba7ddf02"} Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.436395 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" event={"ID":"90fb1eef-5754-4823-8fbc-693ebeb8deb9","Type":"ContainerStarted","Data":"1eb6f1269fd23825d648b77394fed9e3375b2e99bf644d642024294f3d273770"} Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.441565 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84fc97847d-hvzvz" event={"ID":"edd199bd-36e4-4f32-a4cf-237d7a88011a","Type":"ContainerStarted","Data":"e2882613115715bdbbcfe0e1578ce726f508c3cef05a9f763262a8b5670bee9d"} Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.451505 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.458807 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7fdd5dd964-x7n8q" podUID="0509f0f1-226d-4626-998d-b9065a8a634b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.168:8443: connect: connection refused" Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.573944 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:47:53 crc kubenswrapper[4750]: I0309 18:47:53.582990 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b7cc4cf4-w25dt"] Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.594377 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76cfd56bfc-9l7jh" event={"ID":"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0","Type":"ContainerStarted","Data":"dd8ba55221f2cd667260556dbed252c5766fd31db5b763744011aed0ffd4de1b"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.654154 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-64cbf4d6d8-cn7dn"] Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.704023 4750 generic.go:334] "Generic (PLEG): container finished" podID="5be93415-f615-42a7-948f-68829841543e" containerID="0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0" exitCode=0 Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.711841 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b7cc4cf4-w25dt" event={"ID":"150ab1b1-103a-463a-9c5d-fc4231e42674","Type":"ContainerStarted","Data":"c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.711911 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b7cc4cf4-w25dt" event={"ID":"150ab1b1-103a-463a-9c5d-fc4231e42674","Type":"ContainerStarted","Data":"784a25d107b85b36c77b41eb9923fb203e8cd8a0b067cdbc54bef00216fd63d8"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.711930 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" event={"ID":"c44181c7-bb39-45a8-ad11-11f4aab9f8b7","Type":"ContainerStarted","Data":"355272af94253e6578af7d9e42be125da3d3251ac3b02fdd026e62cdbd6ce925"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.711942 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fc5774769-shw5p" event={"ID":"5be93415-f615-42a7-948f-68829841543e","Type":"ContainerDied","Data":"0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.712077 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.731360 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-64cbf4d6d8-cn7dn"] Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.732345 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"c857b2a0-c8c1-452f-99af-a6affa05c8c1","Type":"ContainerStarted","Data":"36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.732420 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"c857b2a0-c8c1-452f-99af-a6affa05c8c1","Type":"ContainerStarted","Data":"424ff5429ce057d88b79873fda82e75df276b23d0d0646f94f9628df631e8d9c"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.761773 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92279388-e576-45e4-a4e6-028db5103301-logs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.761883 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-internal-tls-certs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.761980 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxwxk\" (UniqueName: \"kubernetes.io/projected/92279388-e576-45e4-a4e6-028db5103301-kube-api-access-qxwxk\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.762000 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-config-data\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.762023 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-scripts\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.762051 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-combined-ca-bundle\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.762069 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-public-tls-certs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.800945 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=3.800907715 podStartE2EDuration="3.800907715s" podCreationTimestamp="2026-03-09 18:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:54.786936545 +0000 UTC m=+1356.129408953" watchObservedRunningTime="2026-03-09 18:47:54.800907715 +0000 UTC m=+1356.143380103" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.804356 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84fc97847d-hvzvz" event={"ID":"edd199bd-36e4-4f32-a4cf-237d7a88011a","Type":"ContainerStarted","Data":"1b0e63691cd35c17ed5ac5184ebd75d053c96523f994a79f33733b5aa0d34778"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.805681 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.805710 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.812587 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6594b67854-xmvs4" event={"ID":"f1171881-c394-4042-b092-a6935ed182df","Type":"ContainerStarted","Data":"cbcab80feff8a91f501826f69f89f67e8e9204d9594c8a308c523cfd1f9c504c"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.813349 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.816221 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-577799fd8d-66jqt" event={"ID":"cc0b5242-afa5-4413-9887-432846908b1d","Type":"ContainerStarted","Data":"0ec5ac43145468c25a544af5584315a53e602273454db53bf15205e6ef0e20b2"} Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.816736 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.816837 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.842745 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-84fc97847d-hvzvz" podStartSLOduration=4.842723 podStartE2EDuration="4.842723s" podCreationTimestamp="2026-03-09 18:47:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:54.832331988 +0000 UTC m=+1356.174804386" watchObservedRunningTime="2026-03-09 18:47:54.842723 +0000 UTC m=+1356.185195398" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.864203 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxwxk\" (UniqueName: \"kubernetes.io/projected/92279388-e576-45e4-a4e6-028db5103301-kube-api-access-qxwxk\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.864273 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-config-data\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.864304 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-scripts\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.864353 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-combined-ca-bundle\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.864389 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-public-tls-certs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.864518 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92279388-e576-45e4-a4e6-028db5103301-logs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.864594 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-internal-tls-certs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.871722 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6594b67854-xmvs4" podStartSLOduration=4.871697167 podStartE2EDuration="4.871697167s" podCreationTimestamp="2026-03-09 18:47:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:54.862296742 +0000 UTC m=+1356.204769130" watchObservedRunningTime="2026-03-09 18:47:54.871697167 +0000 UTC m=+1356.214169555" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.874956 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-config-data\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.875761 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92279388-e576-45e4-a4e6-028db5103301-logs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.899036 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-scripts\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.914816 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-public-tls-certs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.917814 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-internal-tls-certs\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.918496 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxwxk\" (UniqueName: \"kubernetes.io/projected/92279388-e576-45e4-a4e6-028db5103301-kube-api-access-qxwxk\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.919090 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92279388-e576-45e4-a4e6-028db5103301-combined-ca-bundle\") pod \"placement-64cbf4d6d8-cn7dn\" (UID: \"92279388-e576-45e4-a4e6-028db5103301\") " pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.934432 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-577799fd8d-66jqt" podStartSLOduration=3.934408651 podStartE2EDuration="3.934408651s" podCreationTimestamp="2026-03-09 18:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:54.901888528 +0000 UTC m=+1356.244360926" watchObservedRunningTime="2026-03-09 18:47:54.934408651 +0000 UTC m=+1356.276881059" Mar 09 18:47:54 crc kubenswrapper[4750]: I0309 18:47:54.944223 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.548712 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-577799fd8d-66jqt"] Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.609193 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-79b78f4bf4-klzwx"] Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.612389 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.616532 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.616695 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.623617 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b78f4bf4-klzwx"] Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.650733 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-64cbf4d6d8-cn7dn"] Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.694278 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6116d3b9-23b6-44a7-bf39-3b139a2a771e-logs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.694403 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-internal-tls-certs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.694490 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-combined-ca-bundle\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.694517 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-public-tls-certs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.694570 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-config-data-custom\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.694593 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-config-data\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.694649 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvmz8\" (UniqueName: \"kubernetes.io/projected/6116d3b9-23b6-44a7-bf39-3b139a2a771e-kube-api-access-kvmz8\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.799155 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-config-data-custom\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.799215 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-config-data\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.799250 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvmz8\" (UniqueName: \"kubernetes.io/projected/6116d3b9-23b6-44a7-bf39-3b139a2a771e-kube-api-access-kvmz8\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.799303 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6116d3b9-23b6-44a7-bf39-3b139a2a771e-logs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.799349 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-internal-tls-certs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.799426 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-combined-ca-bundle\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.799450 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-public-tls-certs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.800012 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6116d3b9-23b6-44a7-bf39-3b139a2a771e-logs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.806941 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-config-data\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.807286 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-internal-tls-certs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.807589 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-public-tls-certs\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.813602 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-combined-ca-bundle\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.814728 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6116d3b9-23b6-44a7-bf39-3b139a2a771e-config-data-custom\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.842398 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvmz8\" (UniqueName: \"kubernetes.io/projected/6116d3b9-23b6-44a7-bf39-3b139a2a771e-kube-api-access-kvmz8\") pod \"barbican-api-79b78f4bf4-klzwx\" (UID: \"6116d3b9-23b6-44a7-bf39-3b139a2a771e\") " pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.867709 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-577799fd8d-66jqt" event={"ID":"cc0b5242-afa5-4413-9887-432846908b1d","Type":"ContainerStarted","Data":"47e15ac3e177cfbd9d23c95dae780ee1d7ccb6bfda812b39cd04378f43ec053c"} Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.882924 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b7cc4cf4-w25dt" event={"ID":"150ab1b1-103a-463a-9c5d-fc4231e42674","Type":"ContainerStarted","Data":"c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514"} Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.883057 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.883083 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.893333 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fc5774769-shw5p" event={"ID":"5be93415-f615-42a7-948f-68829841543e","Type":"ContainerStarted","Data":"e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead"} Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.893388 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.936886 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.962086 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-79b7cc4cf4-w25dt" podStartSLOduration=4.962054471 podStartE2EDuration="4.962054471s" podCreationTimestamp="2026-03-09 18:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:55.915024893 +0000 UTC m=+1357.257497301" watchObservedRunningTime="2026-03-09 18:47:55.962054471 +0000 UTC m=+1357.304526869" Mar 09 18:47:55 crc kubenswrapper[4750]: I0309 18:47:55.970516 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fc5774769-shw5p" podStartSLOduration=5.97049448 podStartE2EDuration="5.97049448s" podCreationTimestamp="2026-03-09 18:47:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:55.960747905 +0000 UTC m=+1357.303220313" watchObservedRunningTime="2026-03-09 18:47:55.97049448 +0000 UTC m=+1357.312966878" Mar 09 18:47:56 crc kubenswrapper[4750]: I0309 18:47:56.914984 4750 generic.go:334] "Generic (PLEG): container finished" podID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerID="0c4186de4c7462289849baaa47e27468f2c826e928ce6d9991a34baefed76f23" exitCode=1 Mar 09 18:47:56 crc kubenswrapper[4750]: I0309 18:47:56.915055 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"9d32a26f-bd14-41a7-872b-e102ea95f696","Type":"ContainerDied","Data":"0c4186de4c7462289849baaa47e27468f2c826e928ce6d9991a34baefed76f23"} Mar 09 18:47:56 crc kubenswrapper[4750]: I0309 18:47:56.915608 4750 scope.go:117] "RemoveContainer" containerID="793969d3d3e5cab9242146e23eb4c826c91cafe7e607d142e01d40eb07324c7c" Mar 09 18:47:56 crc kubenswrapper[4750]: I0309 18:47:56.915682 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-577799fd8d-66jqt" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api-log" containerID="cri-o://0ec5ac43145468c25a544af5584315a53e602273454db53bf15205e6ef0e20b2" gracePeriod=30 Mar 09 18:47:56 crc kubenswrapper[4750]: I0309 18:47:56.915917 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-577799fd8d-66jqt" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api" containerID="cri-o://47e15ac3e177cfbd9d23c95dae780ee1d7ccb6bfda812b39cd04378f43ec053c" gracePeriod=30 Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.314874 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.415990 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.453368 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-config-data\") pod \"9d32a26f-bd14-41a7-872b-e102ea95f696\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.453423 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-combined-ca-bundle\") pod \"9d32a26f-bd14-41a7-872b-e102ea95f696\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.453464 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lppbg\" (UniqueName: \"kubernetes.io/projected/9d32a26f-bd14-41a7-872b-e102ea95f696-kube-api-access-lppbg\") pod \"9d32a26f-bd14-41a7-872b-e102ea95f696\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.453491 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d32a26f-bd14-41a7-872b-e102ea95f696-logs\") pod \"9d32a26f-bd14-41a7-872b-e102ea95f696\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.453565 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-custom-prometheus-ca\") pod \"9d32a26f-bd14-41a7-872b-e102ea95f696\" (UID: \"9d32a26f-bd14-41a7-872b-e102ea95f696\") " Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.457703 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d32a26f-bd14-41a7-872b-e102ea95f696-logs" (OuterVolumeSpecName: "logs") pod "9d32a26f-bd14-41a7-872b-e102ea95f696" (UID: "9d32a26f-bd14-41a7-872b-e102ea95f696"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.490336 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d32a26f-bd14-41a7-872b-e102ea95f696-kube-api-access-lppbg" (OuterVolumeSpecName: "kube-api-access-lppbg") pod "9d32a26f-bd14-41a7-872b-e102ea95f696" (UID: "9d32a26f-bd14-41a7-872b-e102ea95f696"). InnerVolumeSpecName "kube-api-access-lppbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.563292 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "9d32a26f-bd14-41a7-872b-e102ea95f696" (UID: "9d32a26f-bd14-41a7-872b-e102ea95f696"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.563584 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lppbg\" (UniqueName: \"kubernetes.io/projected/9d32a26f-bd14-41a7-872b-e102ea95f696-kube-api-access-lppbg\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.563611 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d32a26f-bd14-41a7-872b-e102ea95f696-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.563641 4750 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.599844 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d32a26f-bd14-41a7-872b-e102ea95f696" (UID: "9d32a26f-bd14-41a7-872b-e102ea95f696"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.644038 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-config-data" (OuterVolumeSpecName: "config-data") pod "9d32a26f-bd14-41a7-872b-e102ea95f696" (UID: "9d32a26f-bd14-41a7-872b-e102ea95f696"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.666353 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.666384 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d32a26f-bd14-41a7-872b-e102ea95f696-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.822106 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b78f4bf4-klzwx"] Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.942962 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" event={"ID":"c44181c7-bb39-45a8-ad11-11f4aab9f8b7","Type":"ContainerStarted","Data":"60f1d44ac70c247a401ddf971705af9563b343dbfdbd8519d79b1900e97af538"} Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.944956 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"9d32a26f-bd14-41a7-872b-e102ea95f696","Type":"ContainerDied","Data":"448fbabe0dcd93888c64acd03db8f6a301f57c5a438fb54c8e58ea16bd2b3336"} Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.945020 4750 scope.go:117] "RemoveContainer" containerID="0c4186de4c7462289849baaa47e27468f2c826e928ce6d9991a34baefed76f23" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.945174 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.951929 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64cbf4d6d8-cn7dn" event={"ID":"92279388-e576-45e4-a4e6-028db5103301","Type":"ContainerStarted","Data":"8009def29f3707e46d736fa6ea1f2ca9b83340939b2a39bfabc86a06aca37ad8"} Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.955886 4750 generic.go:334] "Generic (PLEG): container finished" podID="cc0b5242-afa5-4413-9887-432846908b1d" containerID="47e15ac3e177cfbd9d23c95dae780ee1d7ccb6bfda812b39cd04378f43ec053c" exitCode=0 Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.955939 4750 generic.go:334] "Generic (PLEG): container finished" podID="cc0b5242-afa5-4413-9887-432846908b1d" containerID="0ec5ac43145468c25a544af5584315a53e602273454db53bf15205e6ef0e20b2" exitCode=143 Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.956008 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-577799fd8d-66jqt" event={"ID":"cc0b5242-afa5-4413-9887-432846908b1d","Type":"ContainerDied","Data":"47e15ac3e177cfbd9d23c95dae780ee1d7ccb6bfda812b39cd04378f43ec053c"} Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.956054 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-577799fd8d-66jqt" event={"ID":"cc0b5242-afa5-4413-9887-432846908b1d","Type":"ContainerDied","Data":"0ec5ac43145468c25a544af5584315a53e602273454db53bf15205e6ef0e20b2"} Mar 09 18:47:57 crc kubenswrapper[4750]: I0309 18:47:57.961931 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b78f4bf4-klzwx" event={"ID":"6116d3b9-23b6-44a7-bf39-3b139a2a771e","Type":"ContainerStarted","Data":"6471251d4c940814cfd8a42dea99dd982578b66e5332a93d98359fc4e5a9c819"} Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.240780 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.291925 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data-custom\") pod \"cc0b5242-afa5-4413-9887-432846908b1d\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.292069 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-combined-ca-bundle\") pod \"cc0b5242-afa5-4413-9887-432846908b1d\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.292089 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0b5242-afa5-4413-9887-432846908b1d-logs\") pod \"cc0b5242-afa5-4413-9887-432846908b1d\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.292322 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zv8r\" (UniqueName: \"kubernetes.io/projected/cc0b5242-afa5-4413-9887-432846908b1d-kube-api-access-7zv8r\") pod \"cc0b5242-afa5-4413-9887-432846908b1d\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.292404 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data\") pod \"cc0b5242-afa5-4413-9887-432846908b1d\" (UID: \"cc0b5242-afa5-4413-9887-432846908b1d\") " Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.293970 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc0b5242-afa5-4413-9887-432846908b1d-logs" (OuterVolumeSpecName: "logs") pod "cc0b5242-afa5-4413-9887-432846908b1d" (UID: "cc0b5242-afa5-4413-9887-432846908b1d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.343518 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cc0b5242-afa5-4413-9887-432846908b1d" (UID: "cc0b5242-afa5-4413-9887-432846908b1d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.350325 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc0b5242-afa5-4413-9887-432846908b1d-kube-api-access-7zv8r" (OuterVolumeSpecName: "kube-api-access-7zv8r") pod "cc0b5242-afa5-4413-9887-432846908b1d" (UID: "cc0b5242-afa5-4413-9887-432846908b1d"). InnerVolumeSpecName "kube-api-access-7zv8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.382757 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.403417 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zv8r\" (UniqueName: \"kubernetes.io/projected/cc0b5242-afa5-4413-9887-432846908b1d-kube-api-access-7zv8r\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.403466 4750 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.403476 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0b5242-afa5-4413-9887-432846908b1d-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.435986 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.475863 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc0b5242-afa5-4413-9887-432846908b1d" (UID: "cc0b5242-afa5-4413-9887-432846908b1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.514378 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.515132 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:58 crc kubenswrapper[4750]: E0309 18:47:58.516875 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerName="watcher-decision-engine" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.516921 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerName="watcher-decision-engine" Mar 09 18:47:58 crc kubenswrapper[4750]: E0309 18:47:58.516948 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerName="watcher-decision-engine" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.516956 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerName="watcher-decision-engine" Mar 09 18:47:58 crc kubenswrapper[4750]: E0309 18:47:58.517143 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api-log" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.517153 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api-log" Mar 09 18:47:58 crc kubenswrapper[4750]: E0309 18:47:58.517213 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.517229 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.518338 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api-log" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.518376 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerName="watcher-decision-engine" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.518392 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" containerName="watcher-decision-engine" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.518409 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0b5242-afa5-4413-9887-432846908b1d" containerName="barbican-api" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.521067 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.524691 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.535830 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.609206 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data" (OuterVolumeSpecName: "config-data") pod "cc0b5242-afa5-4413-9887-432846908b1d" (UID: "cc0b5242-afa5-4413-9887-432846908b1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.616824 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-config-data\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.616902 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6796c4d9-5680-4c22-9ef0-815642b22c3f-logs\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.617029 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.617064 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgqsm\" (UniqueName: \"kubernetes.io/projected/6796c4d9-5680-4c22-9ef0-815642b22c3f-kube-api-access-zgqsm\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.617153 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.617221 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0b5242-afa5-4413-9887-432846908b1d-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.719876 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.720119 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgqsm\" (UniqueName: \"kubernetes.io/projected/6796c4d9-5680-4c22-9ef0-815642b22c3f-kube-api-access-zgqsm\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.720191 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.720216 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-config-data\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.720248 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6796c4d9-5680-4c22-9ef0-815642b22c3f-logs\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.720913 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6796c4d9-5680-4c22-9ef0-815642b22c3f-logs\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.733752 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.737378 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.745334 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-config-data\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.751453 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgqsm\" (UniqueName: \"kubernetes.io/projected/6796c4d9-5680-4c22-9ef0-815642b22c3f-kube-api-access-zgqsm\") pod \"watcher-decision-engine-0\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.982742 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb5b94469-qjhvw" event={"ID":"840a6d14-b199-41f2-8adf-124089a1ec27","Type":"ContainerStarted","Data":"62fd9a93aa810da101a980793045a059d2e02724d0b2ab541e57931519be59c5"} Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.982789 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb5b94469-qjhvw" event={"ID":"840a6d14-b199-41f2-8adf-124089a1ec27","Type":"ContainerStarted","Data":"b295c6bb35cc22328f6e4e52dae597bfd67d3ddfee08632cc82696491045a4c6"} Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.989009 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.991680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b78f4bf4-klzwx" event={"ID":"6116d3b9-23b6-44a7-bf39-3b139a2a771e","Type":"ContainerStarted","Data":"2b3edae9f68bf69dafc7367cd4debdfaf394cf6b8dbf878fe070ee66b56cbe94"} Mar 09 18:47:58 crc kubenswrapper[4750]: I0309 18:47:58.995736 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" event={"ID":"c44181c7-bb39-45a8-ad11-11f4aab9f8b7","Type":"ContainerStarted","Data":"83c34baf4d09eff8d303fe713510737aa6d36192ba38f8b77edc2ca8d4e54ba2"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.006342 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" event={"ID":"90fb1eef-5754-4823-8fbc-693ebeb8deb9","Type":"ContainerStarted","Data":"8d6978716d923473b4bfa7329dc2f1564b2a4063245d6c015dfe2207d816a8e2"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.006394 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" event={"ID":"90fb1eef-5754-4823-8fbc-693ebeb8deb9","Type":"ContainerStarted","Data":"72d4fbe46c2719f7769e8ba3a82a101f2bbbb95877f5ada70827c22beba56a38"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.008081 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-fb5b94469-qjhvw" podStartSLOduration=4.314851899 podStartE2EDuration="9.008045754s" podCreationTimestamp="2026-03-09 18:47:50 +0000 UTC" firstStartedPulling="2026-03-09 18:47:52.562865361 +0000 UTC m=+1353.905337759" lastFinishedPulling="2026-03-09 18:47:57.256059216 +0000 UTC m=+1358.598531614" observedRunningTime="2026-03-09 18:47:59.002088453 +0000 UTC m=+1360.344560861" watchObservedRunningTime="2026-03-09 18:47:59.008045754 +0000 UTC m=+1360.350518142" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.019655 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64cbf4d6d8-cn7dn" event={"ID":"92279388-e576-45e4-a4e6-028db5103301","Type":"ContainerStarted","Data":"1b32039c78fdbd7df944bcbc7575f4b260237241e19a384adaec312dececa59c"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.019712 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64cbf4d6d8-cn7dn" event={"ID":"92279388-e576-45e4-a4e6-028db5103301","Type":"ContainerStarted","Data":"cec0af24bd2ae2d3f128cff61229eec96471f1719d4b038055a9806261487a19"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.020899 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.020932 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.026081 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-577799fd8d-66jqt" event={"ID":"cc0b5242-afa5-4413-9887-432846908b1d","Type":"ContainerDied","Data":"c131d8c28d81b78059647080be8d3146e50b9bbde8c4bdcb43e225ca53e638bb"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.026141 4750 scope.go:117] "RemoveContainer" containerID="47e15ac3e177cfbd9d23c95dae780ee1d7ccb6bfda812b39cd04378f43ec053c" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.026286 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-577799fd8d-66jqt" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.041823 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-54d9b479b6-f2tns" podStartSLOduration=4.204475409 podStartE2EDuration="8.041802292s" podCreationTimestamp="2026-03-09 18:47:51 +0000 UTC" firstStartedPulling="2026-03-09 18:47:53.420491141 +0000 UTC m=+1354.762963529" lastFinishedPulling="2026-03-09 18:47:57.257818014 +0000 UTC m=+1358.600290412" observedRunningTime="2026-03-09 18:47:59.030102264 +0000 UTC m=+1360.372574662" watchObservedRunningTime="2026-03-09 18:47:59.041802292 +0000 UTC m=+1360.384274690" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.068579 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76cfd56bfc-9l7jh" event={"ID":"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0","Type":"ContainerStarted","Data":"b7b9383198e76043cb3f1b657fcdc5a838827187b3049ba04fbc40d7757ca513"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.068650 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76cfd56bfc-9l7jh" event={"ID":"968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0","Type":"ContainerStarted","Data":"7afebd0a0ef241190d23b130b7931d3c6ccce46ef64cf66762c6343525586fad"} Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.092484 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-66cf65f968-kqpdg"] Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.095036 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-64cbf4d6d8-cn7dn" podStartSLOduration=5.095023917 podStartE2EDuration="5.095023917s" podCreationTimestamp="2026-03-09 18:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:47:59.060908421 +0000 UTC m=+1360.403380829" watchObservedRunningTime="2026-03-09 18:47:59.095023917 +0000 UTC m=+1360.437496315" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.107657 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" podStartSLOduration=4.675751134 podStartE2EDuration="9.10761802s" podCreationTimestamp="2026-03-09 18:47:50 +0000 UTC" firstStartedPulling="2026-03-09 18:47:52.825337222 +0000 UTC m=+1354.167809620" lastFinishedPulling="2026-03-09 18:47:57.257204108 +0000 UTC m=+1358.599676506" observedRunningTime="2026-03-09 18:47:59.095374018 +0000 UTC m=+1360.437846416" watchObservedRunningTime="2026-03-09 18:47:59.10761802 +0000 UTC m=+1360.450090418" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.121092 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76cfd56bfc-9l7jh" podStartSLOduration=4.318049884 podStartE2EDuration="8.121065055s" podCreationTimestamp="2026-03-09 18:47:51 +0000 UTC" firstStartedPulling="2026-03-09 18:47:53.429741572 +0000 UTC m=+1354.772213970" lastFinishedPulling="2026-03-09 18:47:57.232756743 +0000 UTC m=+1358.575229141" observedRunningTime="2026-03-09 18:47:59.117847948 +0000 UTC m=+1360.460320356" watchObservedRunningTime="2026-03-09 18:47:59.121065055 +0000 UTC m=+1360.463537453" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.185703 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-fb5b94469-qjhvw"] Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.190795 4750 scope.go:117] "RemoveContainer" containerID="0ec5ac43145468c25a544af5584315a53e602273454db53bf15205e6ef0e20b2" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.203356 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-577799fd8d-66jqt"] Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.212451 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-577799fd8d-66jqt"] Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.412315 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d32a26f-bd14-41a7-872b-e102ea95f696" path="/var/lib/kubelet/pods/9d32a26f-bd14-41a7-872b-e102ea95f696/volumes" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.432774 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0b5242-afa5-4413-9887-432846908b1d" path="/var/lib/kubelet/pods/cc0b5242-afa5-4413-9887-432846908b1d/volumes" Mar 09 18:47:59 crc kubenswrapper[4750]: I0309 18:47:59.806657 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.078479 4750 generic.go:334] "Generic (PLEG): container finished" podID="fec39559-ddae-4e19-8941-0fc3209084cd" containerID="fa00c999ec13985a6d12a0ee552edbc27058d1f0a0de89d2e82c9c53574c40f7" exitCode=0 Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.078547 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-89k9z" event={"ID":"fec39559-ddae-4e19-8941-0fc3209084cd","Type":"ContainerDied","Data":"fa00c999ec13985a6d12a0ee552edbc27058d1f0a0de89d2e82c9c53574c40f7"} Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.080995 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b78f4bf4-klzwx" event={"ID":"6116d3b9-23b6-44a7-bf39-3b139a2a771e","Type":"ContainerStarted","Data":"e9d7f1ecc66a99844467cc2af939d833c83d1d13e5552cf5cdfb498d064f52c7"} Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.081428 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.081442 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.093328 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerStarted","Data":"bc6a490fec8c24319e9b4a56ade1c502d219edbe8c0c2af72a71fc412d4a4c9d"} Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.131494 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-79b78f4bf4-klzwx" podStartSLOduration=5.131467197 podStartE2EDuration="5.131467197s" podCreationTimestamp="2026-03-09 18:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:00.126234614 +0000 UTC m=+1361.468707012" watchObservedRunningTime="2026-03-09 18:48:00.131467197 +0000 UTC m=+1361.473939595" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.167001 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551368-ngvhp"] Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.168250 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.177733 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551368-ngvhp"] Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.178375 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.178580 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.179869 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.259832 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qstlf\" (UniqueName: \"kubernetes.io/projected/d81c74f1-ee19-4573-8582-e2fdf6373472-kube-api-access-qstlf\") pod \"auto-csr-approver-29551368-ngvhp\" (UID: \"d81c74f1-ee19-4573-8582-e2fdf6373472\") " pod="openshift-infra/auto-csr-approver-29551368-ngvhp" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.362531 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qstlf\" (UniqueName: \"kubernetes.io/projected/d81c74f1-ee19-4573-8582-e2fdf6373472-kube-api-access-qstlf\") pod \"auto-csr-approver-29551368-ngvhp\" (UID: \"d81c74f1-ee19-4573-8582-e2fdf6373472\") " pod="openshift-infra/auto-csr-approver-29551368-ngvhp" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.385730 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qstlf\" (UniqueName: \"kubernetes.io/projected/d81c74f1-ee19-4573-8582-e2fdf6373472-kube-api-access-qstlf\") pod \"auto-csr-approver-29551368-ngvhp\" (UID: \"d81c74f1-ee19-4573-8582-e2fdf6373472\") " pod="openshift-infra/auto-csr-approver-29551368-ngvhp" Mar 09 18:48:00 crc kubenswrapper[4750]: I0309 18:48:00.492394 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.123129 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-fb5b94469-qjhvw" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker-log" containerID="cri-o://b295c6bb35cc22328f6e4e52dae597bfd67d3ddfee08632cc82696491045a4c6" gracePeriod=30 Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.123505 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-fb5b94469-qjhvw" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker" containerID="cri-o://62fd9a93aa810da101a980793045a059d2e02724d0b2ab541e57931519be59c5" gracePeriod=30 Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.123348 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerStarted","Data":"d8c079ce8ce3061f0b61e824da8fb16b3f717f252681239210602849fef7fded"} Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.124033 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener-log" containerID="cri-o://72d4fbe46c2719f7769e8ba3a82a101f2bbbb95877f5ada70827c22beba56a38" gracePeriod=30 Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.124071 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener" containerID="cri-o://8d6978716d923473b4bfa7329dc2f1564b2a4063245d6c015dfe2207d816a8e2" gracePeriod=30 Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.157175 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=3.157148202 podStartE2EDuration="3.157148202s" podCreationTimestamp="2026-03-09 18:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:01.149955507 +0000 UTC m=+1362.492427905" watchObservedRunningTime="2026-03-09 18:48:01.157148202 +0000 UTC m=+1362.499620600" Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.647856 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.713756 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb69d7559-cxkgc"] Mar 09 18:48:01 crc kubenswrapper[4750]: I0309 18:48:01.714237 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerName="dnsmasq-dns" containerID="cri-o://c3d6fa510aa940a1bd25fc30b24a26fc38dcbd0e75626c836c87f55051ab4d09" gracePeriod=10 Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.133243 4750 generic.go:334] "Generic (PLEG): container finished" podID="840a6d14-b199-41f2-8adf-124089a1ec27" containerID="62fd9a93aa810da101a980793045a059d2e02724d0b2ab541e57931519be59c5" exitCode=0 Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.133280 4750 generic.go:334] "Generic (PLEG): container finished" podID="840a6d14-b199-41f2-8adf-124089a1ec27" containerID="b295c6bb35cc22328f6e4e52dae597bfd67d3ddfee08632cc82696491045a4c6" exitCode=143 Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.133299 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb5b94469-qjhvw" event={"ID":"840a6d14-b199-41f2-8adf-124089a1ec27","Type":"ContainerDied","Data":"62fd9a93aa810da101a980793045a059d2e02724d0b2ab541e57931519be59c5"} Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.133350 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb5b94469-qjhvw" event={"ID":"840a6d14-b199-41f2-8adf-124089a1ec27","Type":"ContainerDied","Data":"b295c6bb35cc22328f6e4e52dae597bfd67d3ddfee08632cc82696491045a4c6"} Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.136032 4750 generic.go:334] "Generic (PLEG): container finished" podID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerID="8d6978716d923473b4bfa7329dc2f1564b2a4063245d6c015dfe2207d816a8e2" exitCode=0 Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.136056 4750 generic.go:334] "Generic (PLEG): container finished" podID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerID="72d4fbe46c2719f7769e8ba3a82a101f2bbbb95877f5ada70827c22beba56a38" exitCode=143 Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.136102 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" event={"ID":"90fb1eef-5754-4823-8fbc-693ebeb8deb9","Type":"ContainerDied","Data":"8d6978716d923473b4bfa7329dc2f1564b2a4063245d6c015dfe2207d816a8e2"} Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.136127 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" event={"ID":"90fb1eef-5754-4823-8fbc-693ebeb8deb9","Type":"ContainerDied","Data":"72d4fbe46c2719f7769e8ba3a82a101f2bbbb95877f5ada70827c22beba56a38"} Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.138917 4750 generic.go:334] "Generic (PLEG): container finished" podID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerID="c3d6fa510aa940a1bd25fc30b24a26fc38dcbd0e75626c836c87f55051ab4d09" exitCode=0 Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.138960 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" event={"ID":"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e","Type":"ContainerDied","Data":"c3d6fa510aa940a1bd25fc30b24a26fc38dcbd0e75626c836c87f55051ab4d09"} Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.315533 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.391596 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Mar 09 18:48:02 crc kubenswrapper[4750]: I0309 18:48:02.645927 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.172:5353: connect: connection refused" Mar 09 18:48:03 crc kubenswrapper[4750]: I0309 18:48:03.182861 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-79b7cc4cf4-w25dt" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 18:48:03 crc kubenswrapper[4750]: I0309 18:48:03.197330 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Mar 09 18:48:03 crc kubenswrapper[4750]: I0309 18:48:03.927065 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:48:04 crc kubenswrapper[4750]: I0309 18:48:04.098993 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:48:04 crc kubenswrapper[4750]: I0309 18:48:04.177977 4750 generic.go:334] "Generic (PLEG): container finished" podID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerID="d8c079ce8ce3061f0b61e824da8fb16b3f717f252681239210602849fef7fded" exitCode=1 Mar 09 18:48:04 crc kubenswrapper[4750]: I0309 18:48:04.178963 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerDied","Data":"d8c079ce8ce3061f0b61e824da8fb16b3f717f252681239210602849fef7fded"} Mar 09 18:48:04 crc kubenswrapper[4750]: I0309 18:48:04.179276 4750 scope.go:117] "RemoveContainer" containerID="d8c079ce8ce3061f0b61e824da8fb16b3f717f252681239210602849fef7fded" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.690114 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.742443 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-89k9z" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.786406 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-nb\") pod \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.786477 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-config-data\") pod \"fec39559-ddae-4e19-8941-0fc3209084cd\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.786521 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-scripts\") pod \"fec39559-ddae-4e19-8941-0fc3209084cd\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787131 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-swift-storage-0\") pod \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787293 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-config\") pod \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787368 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-svc\") pod \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787426 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chvrh\" (UniqueName: \"kubernetes.io/projected/fec39559-ddae-4e19-8941-0fc3209084cd-kube-api-access-chvrh\") pod \"fec39559-ddae-4e19-8941-0fc3209084cd\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787469 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fec39559-ddae-4e19-8941-0fc3209084cd-etc-machine-id\") pod \"fec39559-ddae-4e19-8941-0fc3209084cd\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787504 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-db-sync-config-data\") pod \"fec39559-ddae-4e19-8941-0fc3209084cd\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787561 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-sb\") pod \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787602 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wklpd\" (UniqueName: \"kubernetes.io/projected/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-kube-api-access-wklpd\") pod \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\" (UID: \"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787650 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-combined-ca-bundle\") pod \"fec39559-ddae-4e19-8941-0fc3209084cd\" (UID: \"fec39559-ddae-4e19-8941-0fc3209084cd\") " Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.787846 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fec39559-ddae-4e19-8941-0fc3209084cd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fec39559-ddae-4e19-8941-0fc3209084cd" (UID: "fec39559-ddae-4e19-8941-0fc3209084cd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.789215 4750 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fec39559-ddae-4e19-8941-0fc3209084cd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.804559 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fec39559-ddae-4e19-8941-0fc3209084cd" (UID: "fec39559-ddae-4e19-8941-0fc3209084cd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.804701 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-scripts" (OuterVolumeSpecName: "scripts") pod "fec39559-ddae-4e19-8941-0fc3209084cd" (UID: "fec39559-ddae-4e19-8941-0fc3209084cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.834914 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fec39559-ddae-4e19-8941-0fc3209084cd-kube-api-access-chvrh" (OuterVolumeSpecName: "kube-api-access-chvrh") pod "fec39559-ddae-4e19-8941-0fc3209084cd" (UID: "fec39559-ddae-4e19-8941-0fc3209084cd"). InnerVolumeSpecName "kube-api-access-chvrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.842782 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-kube-api-access-wklpd" (OuterVolumeSpecName: "kube-api-access-wklpd") pod "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" (UID: "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e"). InnerVolumeSpecName "kube-api-access-wklpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.891507 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chvrh\" (UniqueName: \"kubernetes.io/projected/fec39559-ddae-4e19-8941-0fc3209084cd-kube-api-access-chvrh\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.891543 4750 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.891554 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wklpd\" (UniqueName: \"kubernetes.io/projected/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-kube-api-access-wklpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.891563 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.943601 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fec39559-ddae-4e19-8941-0fc3209084cd" (UID: "fec39559-ddae-4e19-8941-0fc3209084cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.978282 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-config" (OuterVolumeSpecName: "config") pod "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" (UID: "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.994050 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:05 crc kubenswrapper[4750]: I0309 18:48:05.994092 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.003485 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" (UID: "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.020382 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" (UID: "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.043901 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" (UID: "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.060194 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" (UID: "c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.092759 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-config-data" (OuterVolumeSpecName: "config-data") pod "fec39559-ddae-4e19-8941-0fc3209084cd" (UID: "fec39559-ddae-4e19-8941-0fc3209084cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.096130 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.096160 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.096171 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.096181 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec39559-ddae-4e19-8941-0fc3209084cd-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.096210 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.200657 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-89k9z" event={"ID":"fec39559-ddae-4e19-8941-0fc3209084cd","Type":"ContainerDied","Data":"5fdc9f515d14510136d8dd08d3bc3d6b17f383963220bd57556a0e715ee72f41"} Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.200704 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fdc9f515d14510136d8dd08d3bc3d6b17f383963220bd57556a0e715ee72f41" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.200786 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-89k9z" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.214445 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" event={"ID":"c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e","Type":"ContainerDied","Data":"b8c0fc0c526c386f4ceaab293fe2b74773add67bcb9853549198cf3b69c98ca6"} Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.214502 4750 scope.go:117] "RemoveContainer" containerID="c3d6fa510aa940a1bd25fc30b24a26fc38dcbd0e75626c836c87f55051ab4d09" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.214766 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb69d7559-cxkgc" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.257134 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb69d7559-cxkgc"] Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.265810 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb69d7559-cxkgc"] Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.340665 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.371287 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.522776 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.535993 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605368 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data\") pod \"840a6d14-b199-41f2-8adf-124089a1ec27\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605424 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwrr9\" (UniqueName: \"kubernetes.io/projected/840a6d14-b199-41f2-8adf-124089a1ec27-kube-api-access-dwrr9\") pod \"840a6d14-b199-41f2-8adf-124089a1ec27\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605548 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data\") pod \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605586 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data-custom\") pod \"840a6d14-b199-41f2-8adf-124089a1ec27\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605608 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg5zq\" (UniqueName: \"kubernetes.io/projected/90fb1eef-5754-4823-8fbc-693ebeb8deb9-kube-api-access-jg5zq\") pod \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605672 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90fb1eef-5754-4823-8fbc-693ebeb8deb9-logs\") pod \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605688 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-combined-ca-bundle\") pod \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605761 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data-custom\") pod \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\" (UID: \"90fb1eef-5754-4823-8fbc-693ebeb8deb9\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605777 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840a6d14-b199-41f2-8adf-124089a1ec27-logs\") pod \"840a6d14-b199-41f2-8adf-124089a1ec27\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.605802 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-combined-ca-bundle\") pod \"840a6d14-b199-41f2-8adf-124089a1ec27\" (UID: \"840a6d14-b199-41f2-8adf-124089a1ec27\") " Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.610872 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/840a6d14-b199-41f2-8adf-124089a1ec27-kube-api-access-dwrr9" (OuterVolumeSpecName: "kube-api-access-dwrr9") pod "840a6d14-b199-41f2-8adf-124089a1ec27" (UID: "840a6d14-b199-41f2-8adf-124089a1ec27"). InnerVolumeSpecName "kube-api-access-dwrr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.613060 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90fb1eef-5754-4823-8fbc-693ebeb8deb9-logs" (OuterVolumeSpecName: "logs") pod "90fb1eef-5754-4823-8fbc-693ebeb8deb9" (UID: "90fb1eef-5754-4823-8fbc-693ebeb8deb9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.613268 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/840a6d14-b199-41f2-8adf-124089a1ec27-logs" (OuterVolumeSpecName: "logs") pod "840a6d14-b199-41f2-8adf-124089a1ec27" (UID: "840a6d14-b199-41f2-8adf-124089a1ec27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.615061 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90fb1eef-5754-4823-8fbc-693ebeb8deb9-kube-api-access-jg5zq" (OuterVolumeSpecName: "kube-api-access-jg5zq") pod "90fb1eef-5754-4823-8fbc-693ebeb8deb9" (UID: "90fb1eef-5754-4823-8fbc-693ebeb8deb9"). InnerVolumeSpecName "kube-api-access-jg5zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.620794 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "840a6d14-b199-41f2-8adf-124089a1ec27" (UID: "840a6d14-b199-41f2-8adf-124089a1ec27"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.620843 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "90fb1eef-5754-4823-8fbc-693ebeb8deb9" (UID: "90fb1eef-5754-4823-8fbc-693ebeb8deb9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.648786 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90fb1eef-5754-4823-8fbc-693ebeb8deb9" (UID: "90fb1eef-5754-4823-8fbc-693ebeb8deb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.655441 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "840a6d14-b199-41f2-8adf-124089a1ec27" (UID: "840a6d14-b199-41f2-8adf-124089a1ec27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.664265 4750 scope.go:117] "RemoveContainer" containerID="0320c6c25a7f5817ddc2e5b13113e5077f79155cbc555208882b4c19bb3fbedf" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.678751 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data" (OuterVolumeSpecName: "config-data") pod "90fb1eef-5754-4823-8fbc-693ebeb8deb9" (UID: "90fb1eef-5754-4823-8fbc-693ebeb8deb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707449 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707480 4750 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707494 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg5zq\" (UniqueName: \"kubernetes.io/projected/90fb1eef-5754-4823-8fbc-693ebeb8deb9-kube-api-access-jg5zq\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707503 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90fb1eef-5754-4823-8fbc-693ebeb8deb9-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707512 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707521 4750 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90fb1eef-5754-4823-8fbc-693ebeb8deb9-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707529 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840a6d14-b199-41f2-8adf-124089a1ec27-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707537 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707546 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwrr9\" (UniqueName: \"kubernetes.io/projected/840a6d14-b199-41f2-8adf-124089a1ec27-kube-api-access-dwrr9\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.707790 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data" (OuterVolumeSpecName: "config-data") pod "840a6d14-b199-41f2-8adf-124089a1ec27" (UID: "840a6d14-b199-41f2-8adf-124089a1ec27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.812798 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/840a6d14-b199-41f2-8adf-124089a1ec27-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.985928 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:06 crc kubenswrapper[4750]: E0309 18:48:06.986407 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986428 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker" Mar 09 18:48:06 crc kubenswrapper[4750]: E0309 18:48:06.986440 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener-log" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986447 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener-log" Mar 09 18:48:06 crc kubenswrapper[4750]: E0309 18:48:06.986468 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986475 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener" Mar 09 18:48:06 crc kubenswrapper[4750]: E0309 18:48:06.986488 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerName="init" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986495 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerName="init" Mar 09 18:48:06 crc kubenswrapper[4750]: E0309 18:48:06.986507 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker-log" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986513 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker-log" Mar 09 18:48:06 crc kubenswrapper[4750]: E0309 18:48:06.986525 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fec39559-ddae-4e19-8941-0fc3209084cd" containerName="cinder-db-sync" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986532 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="fec39559-ddae-4e19-8941-0fc3209084cd" containerName="cinder-db-sync" Mar 09 18:48:06 crc kubenswrapper[4750]: E0309 18:48:06.986546 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerName="dnsmasq-dns" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986551 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerName="dnsmasq-dns" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986769 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="fec39559-ddae-4e19-8941-0fc3209084cd" containerName="cinder-db-sync" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986778 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener-log" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986790 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986802 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" containerName="barbican-worker-log" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986809 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" containerName="dnsmasq-dns" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.986819 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" containerName="barbican-keystone-listener" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.987912 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.993909 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.994550 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.994719 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 09 18:48:06 crc kubenswrapper[4750]: I0309 18:48:06.994825 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-x5tpq" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.005379 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6dfd9d95f-cvfsd"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.007076 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.017963 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.018024 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.018052 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-scripts\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.018144 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.018194 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/efd2c1d8-3a05-412b-9854-1086c8002cf4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.018241 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnj7z\" (UniqueName: \"kubernetes.io/projected/efd2c1d8-3a05-412b-9854-1086c8002cf4-kube-api-access-fnj7z\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.039519 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.076603 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dfd9d95f-cvfsd"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.119928 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120003 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-config\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120029 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120048 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-swift-storage-0\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120073 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-scripts\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120126 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120161 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ljvz\" (UniqueName: \"kubernetes.io/projected/74f3d684-e75b-4b86-9086-e2649c75ee9e-kube-api-access-9ljvz\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120190 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/efd2c1d8-3a05-412b-9854-1086c8002cf4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120235 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-svc\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120252 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-sb\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120275 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnj7z\" (UniqueName: \"kubernetes.io/projected/efd2c1d8-3a05-412b-9854-1086c8002cf4-kube-api-access-fnj7z\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.120301 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-nb\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.121375 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/efd2c1d8-3a05-412b-9854-1086c8002cf4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.127731 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.129526 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.134197 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.136048 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-scripts\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.159994 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnj7z\" (UniqueName: \"kubernetes.io/projected/efd2c1d8-3a05-412b-9854-1086c8002cf4-kube-api-access-fnj7z\") pod \"cinder-scheduler-0\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.184037 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.186986 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: E0309 18:48:07.192035 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.193012 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.202031 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.221807 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.221890 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.221953 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ljvz\" (UniqueName: \"kubernetes.io/projected/74f3d684-e75b-4b86-9086-e2649c75ee9e-kube-api-access-9ljvz\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.221991 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222048 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fe67bbc-847f-43b2-878c-13f9b54f10c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222106 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-svc\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222137 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-sb\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222165 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-scripts\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222206 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-nb\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222232 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tpqw\" (UniqueName: \"kubernetes.io/projected/5fe67bbc-847f-43b2-878c-13f9b54f10c6-kube-api-access-6tpqw\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222266 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fe67bbc-847f-43b2-878c-13f9b54f10c6-logs\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222316 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-config\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.222352 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-swift-storage-0\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.227304 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-svc\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.228280 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551368-ngvhp"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.228873 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-config\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.229767 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-swift-storage-0\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.230143 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-nb\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.239783 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-sb\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.241794 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" event={"ID":"d81c74f1-ee19-4573-8582-e2fdf6373472","Type":"ContainerStarted","Data":"aa163805102285b00feb1dab016c3fed1cb8de9b75d08980a521b5f0a8385970"} Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.264162 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fb5b94469-qjhvw" event={"ID":"840a6d14-b199-41f2-8adf-124089a1ec27","Type":"ContainerDied","Data":"4713f798673c0dc6bdcb80a7894a711720f73c75a4c12e69567d2c83767932d1"} Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.264229 4750 scope.go:117] "RemoveContainer" containerID="62fd9a93aa810da101a980793045a059d2e02724d0b2ab541e57931519be59c5" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.264400 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fb5b94469-qjhvw" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.264497 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ljvz\" (UniqueName: \"kubernetes.io/projected/74f3d684-e75b-4b86-9086-e2649c75ee9e-kube-api-access-9ljvz\") pod \"dnsmasq-dns-6dfd9d95f-cvfsd\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.291787 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerStarted","Data":"47db330f58e384d75e7b5a4e8eabdd01fc7f8f80577c7d541f52eb618d160dd7"} Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.291996 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="ceilometer-notification-agent" containerID="cri-o://6e6034f7d85033a58557e0e8af1f961c2ebd2a12c6f963ee1aa5060d9ca53b8e" gracePeriod=30 Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.292084 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.292127 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="proxy-httpd" containerID="cri-o://47db330f58e384d75e7b5a4e8eabdd01fc7f8f80577c7d541f52eb618d160dd7" gracePeriod=30 Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.292169 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="sg-core" containerID="cri-o://5e6269255415c111f968e23ad15921e57205996484e746f79d35fde5ab0b990e" gracePeriod=30 Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.299550 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerStarted","Data":"8a942558352aa440b404d831fbe9e6f7cef4864baa5eef35869bb3c60f919a6a"} Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.314602 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" event={"ID":"90fb1eef-5754-4823-8fbc-693ebeb8deb9","Type":"ContainerDied","Data":"1eb6f1269fd23825d648b77394fed9e3375b2e99bf644d642024294f3d273770"} Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.314717 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66cf65f968-kqpdg" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.321823 4750 scope.go:117] "RemoveContainer" containerID="b295c6bb35cc22328f6e4e52dae597bfd67d3ddfee08632cc82696491045a4c6" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324374 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324429 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324475 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fe67bbc-847f-43b2-878c-13f9b54f10c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324507 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-scripts\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324536 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tpqw\" (UniqueName: \"kubernetes.io/projected/5fe67bbc-847f-43b2-878c-13f9b54f10c6-kube-api-access-6tpqw\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324577 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fe67bbc-847f-43b2-878c-13f9b54f10c6-logs\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324745 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fe67bbc-847f-43b2-878c-13f9b54f10c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.324819 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.328649 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fe67bbc-847f-43b2-878c-13f9b54f10c6-logs\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.329287 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.329736 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.329837 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-fb5b94469-qjhvw"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.335163 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.339311 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-scripts\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.340269 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.353398 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.355708 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tpqw\" (UniqueName: \"kubernetes.io/projected/5fe67bbc-847f-43b2-878c-13f9b54f10c6-kube-api-access-6tpqw\") pod \"cinder-api-0\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.438218 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e" path="/var/lib/kubelet/pods/c55e5e2c-4e0b-4d7b-85e6-9cb49c96993e/volumes" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.439789 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-fb5b94469-qjhvw"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.450746 4750 scope.go:117] "RemoveContainer" containerID="8d6978716d923473b4bfa7329dc2f1564b2a4063245d6c015dfe2207d816a8e2" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.499176 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-66cf65f968-kqpdg"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.513005 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-66cf65f968-kqpdg"] Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.518126 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.519360 4750 scope.go:117] "RemoveContainer" containerID="72d4fbe46c2719f7769e8ba3a82a101f2bbbb95877f5ada70827c22beba56a38" Mar 09 18:48:07 crc kubenswrapper[4750]: I0309 18:48:07.819809 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.085375 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.165591 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dfd9d95f-cvfsd"] Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.208385 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.232385 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-748768969-j7m74"] Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.232726 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-748768969-j7m74" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-api" containerID="cri-o://89b0977a0ec3c9f5642c50c7e3f122d603979f0c03ed7c79afc7d8bf99ce5b5b" gracePeriod=30 Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.234842 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-748768969-j7m74" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-httpd" containerID="cri-o://0d8051df72ad676824e604b83bb7b43891baff0d89ef21f42ae3b5cb0c1ab46b" gracePeriod=30 Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.249890 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cd6d4c84f-zrhsg"] Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.251732 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.269886 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cd6d4c84f-zrhsg"] Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.272172 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-748768969-j7m74" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.174:9696/\": read tcp 10.217.0.2:42240->10.217.0.174:9696: read: connection reset by peer" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.354403 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"efd2c1d8-3a05-412b-9854-1086c8002cf4","Type":"ContainerStarted","Data":"799ba6a5cdd97f2bd8441f4deb6680f8bd1ad86215dffff2bdcdc4ec6cf62b50"} Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.366195 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" event={"ID":"74f3d684-e75b-4b86-9086-e2649c75ee9e","Type":"ContainerStarted","Data":"ad88955e1b64b42e80bf03f2bcaf1222d483a93bc20a3540b662f1dc8613d558"} Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.370871 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerID="5e6269255415c111f968e23ad15921e57205996484e746f79d35fde5ab0b990e" exitCode=2 Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.371008 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerDied","Data":"5e6269255415c111f968e23ad15921e57205996484e746f79d35fde5ab0b990e"} Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.375360 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-httpd-config\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.375434 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-internal-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.375477 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpmnc\" (UniqueName: \"kubernetes.io/projected/cd28fe12-963a-4e2a-be20-6e58f8034a6f-kube-api-access-kpmnc\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.375544 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-ovndb-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.375572 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-config\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.375648 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-public-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.375701 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-combined-ca-bundle\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.384854 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fe67bbc-847f-43b2-878c-13f9b54f10c6","Type":"ContainerStarted","Data":"4af08225eb472430343edcac649fb94efba0180d48ce0a8ea35af161880244c5"} Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.424959 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.478849 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-httpd-config\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.478943 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-internal-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.478982 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpmnc\" (UniqueName: \"kubernetes.io/projected/cd28fe12-963a-4e2a-be20-6e58f8034a6f-kube-api-access-kpmnc\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.479049 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-ovndb-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.479081 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-config\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.479137 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-public-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.479178 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-combined-ca-bundle\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.488504 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-internal-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.490471 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-ovndb-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.490719 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-public-tls-certs\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.492427 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-config\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.493341 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-combined-ca-bundle\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.493363 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cd28fe12-963a-4e2a-be20-6e58f8034a6f-httpd-config\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.500674 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpmnc\" (UniqueName: \"kubernetes.io/projected/cd28fe12-963a-4e2a-be20-6e58f8034a6f-kube-api-access-kpmnc\") pod \"neutron-6cd6d4c84f-zrhsg\" (UID: \"cd28fe12-963a-4e2a-be20-6e58f8034a6f\") " pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.656117 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.747854 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b78f4bf4-klzwx" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.833355 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79b7cc4cf4-w25dt"] Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.844072 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79b7cc4cf4-w25dt" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api-log" containerID="cri-o://c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7" gracePeriod=30 Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.844576 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79b7cc4cf4-w25dt" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api" containerID="cri-o://c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514" gracePeriod=30 Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.880882 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79b7cc4cf4-w25dt" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": EOF" Mar 09 18:48:08 crc kubenswrapper[4750]: I0309 18:48:08.991107 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.059606 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.445932 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="840a6d14-b199-41f2-8adf-124089a1ec27" path="/var/lib/kubelet/pods/840a6d14-b199-41f2-8adf-124089a1ec27/volumes" Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.446577 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90fb1eef-5754-4823-8fbc-693ebeb8deb9" path="/var/lib/kubelet/pods/90fb1eef-5754-4823-8fbc-693ebeb8deb9/volumes" Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.460715 4750 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433] : Timed out while waiting for systemd to remove kubepods-besteffort-pod3b22ce97_a0b4_4e8b_a2ea_9aef5afb9433.slice" Mar 09 18:48:09 crc kubenswrapper[4750]: E0309 18:48:09.460767 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433] : Timed out while waiting for systemd to remove kubepods-besteffort-pod3b22ce97_a0b4_4e8b_a2ea_9aef5afb9433.slice" pod="openstack/watcher-api-0" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.583988 4750 generic.go:334] "Generic (PLEG): container finished" podID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerID="559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c" exitCode=0 Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.584101 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" event={"ID":"74f3d684-e75b-4b86-9086-e2649c75ee9e","Type":"ContainerDied","Data":"559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c"} Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.622159 4750 generic.go:334] "Generic (PLEG): container finished" podID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerID="c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7" exitCode=143 Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.622335 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b7cc4cf4-w25dt" event={"ID":"150ab1b1-103a-463a-9c5d-fc4231e42674","Type":"ContainerDied","Data":"c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7"} Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.693306 4750 generic.go:334] "Generic (PLEG): container finished" podID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerID="0d8051df72ad676824e604b83bb7b43891baff0d89ef21f42ae3b5cb0c1ab46b" exitCode=0 Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.695226 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748768969-j7m74" event={"ID":"72dade1d-7fb7-4118-a665-7b2fdfc56a9e","Type":"ContainerDied","Data":"0d8051df72ad676824e604b83bb7b43891baff0d89ef21f42ae3b5cb0c1ab46b"} Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.695944 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.868846 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.902100 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:09 crc kubenswrapper[4750]: I0309 18:48:09.929258 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7fdd5dd964-x7n8q" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.011693 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cd6d4c84f-zrhsg"] Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.161451 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5ddc4b4b8d-f5q7t"] Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.163165 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon-log" containerID="cri-o://7c882d7feb763fa768bb968ae79a038f49e22c03454de50f1ddfe39b393e082b" gracePeriod=30 Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.163719 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" containerID="cri-o://8f715bad4ddba91fed463c55ef1bb23df16777368f645f8c1a8928028b49eb87" gracePeriod=30 Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.201383 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.167:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.215182 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.167:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.416697 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-748768969-j7m74" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.174:9696/\": dial tcp 10.217.0.174:9696: connect: connection refused" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.735427 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" event={"ID":"d81c74f1-ee19-4573-8582-e2fdf6373472","Type":"ContainerStarted","Data":"85ec18044d65e2b9a602eb3c76782ebc70dd51b5f4e6463deea04d551c4f57a3"} Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.756154 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fe67bbc-847f-43b2-878c-13f9b54f10c6","Type":"ContainerStarted","Data":"3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf"} Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.758925 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"efd2c1d8-3a05-412b-9854-1086c8002cf4","Type":"ContainerStarted","Data":"b2aef219fc788d84813540d7b11774c8bf24d266e97ec717a2d9c1cee4a0e09e"} Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.763029 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" event={"ID":"74f3d684-e75b-4b86-9086-e2649c75ee9e","Type":"ContainerStarted","Data":"8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784"} Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.763076 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.771097 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" podStartSLOduration=9.545816258 podStartE2EDuration="10.771076376s" podCreationTimestamp="2026-03-09 18:48:00 +0000 UTC" firstStartedPulling="2026-03-09 18:48:07.217874291 +0000 UTC m=+1368.560346689" lastFinishedPulling="2026-03-09 18:48:08.443134409 +0000 UTC m=+1369.785606807" observedRunningTime="2026-03-09 18:48:10.752544952 +0000 UTC m=+1372.095017350" watchObservedRunningTime="2026-03-09 18:48:10.771076376 +0000 UTC m=+1372.113548774" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.780858 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerID="6e6034f7d85033a58557e0e8af1f961c2ebd2a12c6f963ee1aa5060d9ca53b8e" exitCode=0 Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.781236 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerDied","Data":"6e6034f7d85033a58557e0e8af1f961c2ebd2a12c6f963ee1aa5060d9ca53b8e"} Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.789185 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.789447 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cd6d4c84f-zrhsg" event={"ID":"cd28fe12-963a-4e2a-be20-6e58f8034a6f","Type":"ContainerStarted","Data":"5dce715c18ea06e54ff2990c440b4c41bdbdbe09b47c1e884df5836aed471eac"} Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.789621 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cd6d4c84f-zrhsg" event={"ID":"cd28fe12-963a-4e2a-be20-6e58f8034a6f","Type":"ContainerStarted","Data":"3c251b326c7f9390af7a1fce295f15db9c04d28a35aa5aee81ad8b7d4088fe37"} Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.798857 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" podStartSLOduration=4.798836869 podStartE2EDuration="4.798836869s" podCreationTimestamp="2026-03-09 18:48:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:10.785801776 +0000 UTC m=+1372.128274174" watchObservedRunningTime="2026-03-09 18:48:10.798836869 +0000 UTC m=+1372.141309267" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.844908 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.862685 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.879733 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.881458 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.885168 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.899004 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.913347 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvb4b\" (UniqueName: \"kubernetes.io/projected/180bc81d-e02f-4754-93cf-e0345d11829d-kube-api-access-vvb4b\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.913432 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/180bc81d-e02f-4754-93cf-e0345d11829d-logs\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.913478 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.913515 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-config-data\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:10 crc kubenswrapper[4750]: I0309 18:48:10.913552 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.015115 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvb4b\" (UniqueName: \"kubernetes.io/projected/180bc81d-e02f-4754-93cf-e0345d11829d-kube-api-access-vvb4b\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.015217 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/180bc81d-e02f-4754-93cf-e0345d11829d-logs\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.015257 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.015296 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-config-data\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.015347 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.016397 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/180bc81d-e02f-4754-93cf-e0345d11829d-logs\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.042732 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.045845 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-config-data\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.049290 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.061486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvb4b\" (UniqueName: \"kubernetes.io/projected/180bc81d-e02f-4754-93cf-e0345d11829d-kube-api-access-vvb4b\") pod \"watcher-api-0\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.240944 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.392989 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433" path="/var/lib/kubelet/pods/3b22ce97-a0b4-4e8b-a2ea-9aef5afb9433/volumes" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.850207 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cd6d4c84f-zrhsg" event={"ID":"cd28fe12-963a-4e2a-be20-6e58f8034a6f","Type":"ContainerStarted","Data":"7bfac309182fa38bd8b6ea680941447760b30e9d16c98f5cc80546d1b76f38a7"} Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.850611 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.855537 4750 generic.go:334] "Generic (PLEG): container finished" podID="d81c74f1-ee19-4573-8582-e2fdf6373472" containerID="85ec18044d65e2b9a602eb3c76782ebc70dd51b5f4e6463deea04d551c4f57a3" exitCode=0 Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.856781 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" event={"ID":"d81c74f1-ee19-4573-8582-e2fdf6373472","Type":"ContainerDied","Data":"85ec18044d65e2b9a602eb3c76782ebc70dd51b5f4e6463deea04d551c4f57a3"} Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.896623 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:11 crc kubenswrapper[4750]: I0309 18:48:11.909730 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cd6d4c84f-zrhsg" podStartSLOduration=3.9097083599999998 podStartE2EDuration="3.90970836s" podCreationTimestamp="2026-03-09 18:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:11.887931418 +0000 UTC m=+1373.230403816" watchObservedRunningTime="2026-03-09 18:48:11.90970836 +0000 UTC m=+1373.252180758" Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.869410 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"180bc81d-e02f-4754-93cf-e0345d11829d","Type":"ContainerStarted","Data":"4aa97216ea9a86b28cadd4845f784fe4f9363054e5dded4e036d36e36dbc706c"} Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.870230 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.870282 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"180bc81d-e02f-4754-93cf-e0345d11829d","Type":"ContainerStarted","Data":"4bb993426867ef727ad76e8b4ef6a9d448bb3a2b6834c1bde63f1c8a968deda4"} Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.870317 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"180bc81d-e02f-4754-93cf-e0345d11829d","Type":"ContainerStarted","Data":"3e897e6ae1c06521196e8de8fae3d284a1d7437623b6519e2e1faac6a014ee52"} Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.872333 4750 generic.go:334] "Generic (PLEG): container finished" podID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerID="8f715bad4ddba91fed463c55ef1bb23df16777368f645f8c1a8928028b49eb87" exitCode=0 Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.872421 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddc4b4b8d-f5q7t" event={"ID":"ef45ec45-1ef2-46e6-bf33-0c092885679f","Type":"ContainerDied","Data":"8f715bad4ddba91fed463c55ef1bb23df16777368f645f8c1a8928028b49eb87"} Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.875518 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fe67bbc-847f-43b2-878c-13f9b54f10c6","Type":"ContainerStarted","Data":"cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492"} Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.875603 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api-log" containerID="cri-o://3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf" gracePeriod=30 Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.875676 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api" containerID="cri-o://cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492" gracePeriod=30 Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.875737 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.884266 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"efd2c1d8-3a05-412b-9854-1086c8002cf4","Type":"ContainerStarted","Data":"846929547db069fe8adbbcf0a02d640976c07c15d2e9d7eb9ef85fd81f1a0aa7"} Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.891006 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=2.89097956 podStartE2EDuration="2.89097956s" podCreationTimestamp="2026-03-09 18:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:12.889338305 +0000 UTC m=+1374.231810703" watchObservedRunningTime="2026-03-09 18:48:12.89097956 +0000 UTC m=+1374.233451958" Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.923074 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.033596316 podStartE2EDuration="6.923049801s" podCreationTimestamp="2026-03-09 18:48:06 +0000 UTC" firstStartedPulling="2026-03-09 18:48:08.063396853 +0000 UTC m=+1369.405869251" lastFinishedPulling="2026-03-09 18:48:08.952850338 +0000 UTC m=+1370.295322736" observedRunningTime="2026-03-09 18:48:12.913914773 +0000 UTC m=+1374.256387171" watchObservedRunningTime="2026-03-09 18:48:12.923049801 +0000 UTC m=+1374.265522199" Mar 09 18:48:12 crc kubenswrapper[4750]: I0309 18:48:12.992670 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.992623131 podStartE2EDuration="5.992623131s" podCreationTimestamp="2026-03-09 18:48:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:12.946339223 +0000 UTC m=+1374.288811631" watchObservedRunningTime="2026-03-09 18:48:12.992623131 +0000 UTC m=+1374.335095529" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.356005 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.167:8443: connect: connection refused" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.566936 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.703502 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qstlf\" (UniqueName: \"kubernetes.io/projected/d81c74f1-ee19-4573-8582-e2fdf6373472-kube-api-access-qstlf\") pod \"d81c74f1-ee19-4573-8582-e2fdf6373472\" (UID: \"d81c74f1-ee19-4573-8582-e2fdf6373472\") " Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.710376 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d81c74f1-ee19-4573-8582-e2fdf6373472-kube-api-access-qstlf" (OuterVolumeSpecName: "kube-api-access-qstlf") pod "d81c74f1-ee19-4573-8582-e2fdf6373472" (UID: "d81c74f1-ee19-4573-8582-e2fdf6373472"). InnerVolumeSpecName "kube-api-access-qstlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.807278 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qstlf\" (UniqueName: \"kubernetes.io/projected/d81c74f1-ee19-4573-8582-e2fdf6373472-kube-api-access-qstlf\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.872210 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551362-kgfhj"] Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.884834 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551362-kgfhj"] Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.943345 4750 generic.go:334] "Generic (PLEG): container finished" podID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerID="8a942558352aa440b404d831fbe9e6f7cef4864baa5eef35869bb3c60f919a6a" exitCode=1 Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.943461 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerDied","Data":"8a942558352aa440b404d831fbe9e6f7cef4864baa5eef35869bb3c60f919a6a"} Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.943528 4750 scope.go:117] "RemoveContainer" containerID="d8c079ce8ce3061f0b61e824da8fb16b3f717f252681239210602849fef7fded" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.946469 4750 scope.go:117] "RemoveContainer" containerID="8a942558352aa440b404d831fbe9e6f7cef4864baa5eef35869bb3c60f919a6a" Mar 09 18:48:13 crc kubenswrapper[4750]: E0309 18:48:13.946930 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(6796c4d9-5680-4c22-9ef0-815642b22c3f)\"" pod="openstack/watcher-decision-engine-0" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.987200 4750 generic.go:334] "Generic (PLEG): container finished" podID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerID="89b0977a0ec3c9f5642c50c7e3f122d603979f0c03ed7c79afc7d8bf99ce5b5b" exitCode=0 Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.987352 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748768969-j7m74" event={"ID":"72dade1d-7fb7-4118-a665-7b2fdfc56a9e","Type":"ContainerDied","Data":"89b0977a0ec3c9f5642c50c7e3f122d603979f0c03ed7c79afc7d8bf99ce5b5b"} Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.998067 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" event={"ID":"d81c74f1-ee19-4573-8582-e2fdf6373472","Type":"ContainerDied","Data":"aa163805102285b00feb1dab016c3fed1cb8de9b75d08980a521b5f0a8385970"} Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.998109 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa163805102285b00feb1dab016c3fed1cb8de9b75d08980a521b5f0a8385970" Mar 09 18:48:13 crc kubenswrapper[4750]: I0309 18:48:13.998184 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551368-ngvhp" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.021254 4750 generic.go:334] "Generic (PLEG): container finished" podID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerID="3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf" exitCode=143 Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.025716 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fe67bbc-847f-43b2-878c-13f9b54f10c6","Type":"ContainerDied","Data":"3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf"} Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.127536 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748768969-j7m74" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.216563 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-combined-ca-bundle\") pod \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.216975 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-internal-tls-certs\") pod \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.217058 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-public-tls-certs\") pod \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.217074 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-config\") pod \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.217095 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-ovndb-tls-certs\") pod \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.217167 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-httpd-config\") pod \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.217211 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfhxl\" (UniqueName: \"kubernetes.io/projected/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-kube-api-access-jfhxl\") pod \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\" (UID: \"72dade1d-7fb7-4118-a665-7b2fdfc56a9e\") " Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.235806 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "72dade1d-7fb7-4118-a665-7b2fdfc56a9e" (UID: "72dade1d-7fb7-4118-a665-7b2fdfc56a9e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.245066 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-kube-api-access-jfhxl" (OuterVolumeSpecName: "kube-api-access-jfhxl") pod "72dade1d-7fb7-4118-a665-7b2fdfc56a9e" (UID: "72dade1d-7fb7-4118-a665-7b2fdfc56a9e"). InnerVolumeSpecName "kube-api-access-jfhxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.320181 4750 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.320215 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfhxl\" (UniqueName: \"kubernetes.io/projected/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-kube-api-access-jfhxl\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.330939 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72dade1d-7fb7-4118-a665-7b2fdfc56a9e" (UID: "72dade1d-7fb7-4118-a665-7b2fdfc56a9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.348528 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "72dade1d-7fb7-4118-a665-7b2fdfc56a9e" (UID: "72dade1d-7fb7-4118-a665-7b2fdfc56a9e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.352797 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "72dade1d-7fb7-4118-a665-7b2fdfc56a9e" (UID: "72dade1d-7fb7-4118-a665-7b2fdfc56a9e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.399463 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-config" (OuterVolumeSpecName: "config") pod "72dade1d-7fb7-4118-a665-7b2fdfc56a9e" (UID: "72dade1d-7fb7-4118-a665-7b2fdfc56a9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.418786 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "72dade1d-7fb7-4118-a665-7b2fdfc56a9e" (UID: "72dade1d-7fb7-4118-a665-7b2fdfc56a9e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.422037 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.422061 4750 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.422070 4750 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.422080 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:14 crc kubenswrapper[4750]: I0309 18:48:14.422092 4750 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72dade1d-7fb7-4118-a665-7b2fdfc56a9e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.031926 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-748768969-j7m74" event={"ID":"72dade1d-7fb7-4118-a665-7b2fdfc56a9e","Type":"ContainerDied","Data":"84c5cffc803cc82228e944e18adf5d58dbee69b2e90b171697d30a95e48b5917"} Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.031966 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-748768969-j7m74" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.031974 4750 scope.go:117] "RemoveContainer" containerID="0d8051df72ad676824e604b83bb7b43891baff0d89ef21f42ae3b5cb0c1ab46b" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.057547 4750 scope.go:117] "RemoveContainer" containerID="89b0977a0ec3c9f5642c50c7e3f122d603979f0c03ed7c79afc7d8bf99ce5b5b" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.083040 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-748768969-j7m74"] Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.091839 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-748768969-j7m74"] Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.323253 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79b7cc4cf4-w25dt" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": read tcp 10.217.0.2:36532->10.217.0.184:9311: read: connection reset by peer" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.323265 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79b7cc4cf4-w25dt" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": read tcp 10.217.0.2:36548->10.217.0.184:9311: read: connection reset by peer" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.416428 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" path="/var/lib/kubelet/pods/72dade1d-7fb7-4118-a665-7b2fdfc56a9e/volumes" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.417089 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b622dcef-e250-4105-bb3c-245620f16c39" path="/var/lib/kubelet/pods/b622dcef-e250-4105-bb3c-245620f16c39/volumes" Mar 09 18:48:15 crc kubenswrapper[4750]: I0309 18:48:15.980744 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.049524 4750 generic.go:334] "Generic (PLEG): container finished" podID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerID="c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514" exitCode=0 Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.049599 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b7cc4cf4-w25dt" event={"ID":"150ab1b1-103a-463a-9c5d-fc4231e42674","Type":"ContainerDied","Data":"c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514"} Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.049821 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b7cc4cf4-w25dt" event={"ID":"150ab1b1-103a-463a-9c5d-fc4231e42674","Type":"ContainerDied","Data":"784a25d107b85b36c77b41eb9923fb203e8cd8a0b067cdbc54bef00216fd63d8"} Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.049840 4750 scope.go:117] "RemoveContainer" containerID="c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.049618 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b7cc4cf4-w25dt" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.054889 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/150ab1b1-103a-463a-9c5d-fc4231e42674-logs\") pod \"150ab1b1-103a-463a-9c5d-fc4231e42674\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.054941 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-combined-ca-bundle\") pod \"150ab1b1-103a-463a-9c5d-fc4231e42674\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.054973 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data-custom\") pod \"150ab1b1-103a-463a-9c5d-fc4231e42674\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.055047 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data\") pod \"150ab1b1-103a-463a-9c5d-fc4231e42674\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.055194 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v2xw\" (UniqueName: \"kubernetes.io/projected/150ab1b1-103a-463a-9c5d-fc4231e42674-kube-api-access-4v2xw\") pod \"150ab1b1-103a-463a-9c5d-fc4231e42674\" (UID: \"150ab1b1-103a-463a-9c5d-fc4231e42674\") " Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.056193 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/150ab1b1-103a-463a-9c5d-fc4231e42674-logs" (OuterVolumeSpecName: "logs") pod "150ab1b1-103a-463a-9c5d-fc4231e42674" (UID: "150ab1b1-103a-463a-9c5d-fc4231e42674"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.056351 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/150ab1b1-103a-463a-9c5d-fc4231e42674-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.066534 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/150ab1b1-103a-463a-9c5d-fc4231e42674-kube-api-access-4v2xw" (OuterVolumeSpecName: "kube-api-access-4v2xw") pod "150ab1b1-103a-463a-9c5d-fc4231e42674" (UID: "150ab1b1-103a-463a-9c5d-fc4231e42674"). InnerVolumeSpecName "kube-api-access-4v2xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.066847 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "150ab1b1-103a-463a-9c5d-fc4231e42674" (UID: "150ab1b1-103a-463a-9c5d-fc4231e42674"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.093064 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "150ab1b1-103a-463a-9c5d-fc4231e42674" (UID: "150ab1b1-103a-463a-9c5d-fc4231e42674"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.095690 4750 scope.go:117] "RemoveContainer" containerID="c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.119574 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data" (OuterVolumeSpecName: "config-data") pod "150ab1b1-103a-463a-9c5d-fc4231e42674" (UID: "150ab1b1-103a-463a-9c5d-fc4231e42674"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.159179 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v2xw\" (UniqueName: \"kubernetes.io/projected/150ab1b1-103a-463a-9c5d-fc4231e42674-kube-api-access-4v2xw\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.159213 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.159222 4750 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.159232 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/150ab1b1-103a-463a-9c5d-fc4231e42674-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.232572 4750 scope.go:117] "RemoveContainer" containerID="c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514" Mar 09 18:48:16 crc kubenswrapper[4750]: E0309 18:48:16.233085 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514\": container with ID starting with c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514 not found: ID does not exist" containerID="c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.233138 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514"} err="failed to get container status \"c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514\": rpc error: code = NotFound desc = could not find container \"c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514\": container with ID starting with c0ef284574575b4fadb4aa572ace36032f43994fe724c46dc633e4e5edd03514 not found: ID does not exist" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.233172 4750 scope.go:117] "RemoveContainer" containerID="c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7" Mar 09 18:48:16 crc kubenswrapper[4750]: E0309 18:48:16.233518 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7\": container with ID starting with c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7 not found: ID does not exist" containerID="c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.233558 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7"} err="failed to get container status \"c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7\": rpc error: code = NotFound desc = could not find container \"c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7\": container with ID starting with c9495b10677a1f5909a1f949ef5f1d4d142a55bd89dd593386dd4f62eb9d45c7 not found: ID does not exist" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.241297 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.241408 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.398470 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79b7cc4cf4-w25dt"] Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.409105 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-79b7cc4cf4-w25dt"] Mar 09 18:48:16 crc kubenswrapper[4750]: I0309 18:48:16.438087 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Mar 09 18:48:17 crc kubenswrapper[4750]: I0309 18:48:17.331857 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 18:48:17 crc kubenswrapper[4750]: I0309 18:48:17.337728 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:48:17 crc kubenswrapper[4750]: I0309 18:48:17.384235 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" path="/var/lib/kubelet/pods/150ab1b1-103a-463a-9c5d-fc4231e42674/volumes" Mar 09 18:48:17 crc kubenswrapper[4750]: I0309 18:48:17.417125 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fc5774769-shw5p"] Mar 09 18:48:17 crc kubenswrapper[4750]: I0309 18:48:17.417369 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fc5774769-shw5p" podUID="5be93415-f615-42a7-948f-68829841543e" containerName="dnsmasq-dns" containerID="cri-o://e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead" gracePeriod=10 Mar 09 18:48:17 crc kubenswrapper[4750]: I0309 18:48:17.559094 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 09 18:48:17 crc kubenswrapper[4750]: I0309 18:48:17.983660 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.017150 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-config\") pod \"5be93415-f615-42a7-948f-68829841543e\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.017245 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-sb\") pod \"5be93415-f615-42a7-948f-68829841543e\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.017430 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snxst\" (UniqueName: \"kubernetes.io/projected/5be93415-f615-42a7-948f-68829841543e-kube-api-access-snxst\") pod \"5be93415-f615-42a7-948f-68829841543e\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.017567 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-nb\") pod \"5be93415-f615-42a7-948f-68829841543e\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.017595 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-swift-storage-0\") pod \"5be93415-f615-42a7-948f-68829841543e\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.017618 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-svc\") pod \"5be93415-f615-42a7-948f-68829841543e\" (UID: \"5be93415-f615-42a7-948f-68829841543e\") " Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.041797 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5be93415-f615-42a7-948f-68829841543e-kube-api-access-snxst" (OuterVolumeSpecName: "kube-api-access-snxst") pod "5be93415-f615-42a7-948f-68829841543e" (UID: "5be93415-f615-42a7-948f-68829841543e"). InnerVolumeSpecName "kube-api-access-snxst". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.072711 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5be93415-f615-42a7-948f-68829841543e" (UID: "5be93415-f615-42a7-948f-68829841543e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.081599 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5be93415-f615-42a7-948f-68829841543e" (UID: "5be93415-f615-42a7-948f-68829841543e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.092796 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5be93415-f615-42a7-948f-68829841543e" (UID: "5be93415-f615-42a7-948f-68829841543e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.121082 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snxst\" (UniqueName: \"kubernetes.io/projected/5be93415-f615-42a7-948f-68829841543e-kube-api-access-snxst\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.121131 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.121146 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.121159 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.130514 4750 generic.go:334] "Generic (PLEG): container finished" podID="5be93415-f615-42a7-948f-68829841543e" containerID="e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead" exitCode=0 Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.131654 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-config" (OuterVolumeSpecName: "config") pod "5be93415-f615-42a7-948f-68829841543e" (UID: "5be93415-f615-42a7-948f-68829841543e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.131752 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fc5774769-shw5p" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.132046 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fc5774769-shw5p" event={"ID":"5be93415-f615-42a7-948f-68829841543e","Type":"ContainerDied","Data":"e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead"} Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.132271 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fc5774769-shw5p" event={"ID":"5be93415-f615-42a7-948f-68829841543e","Type":"ContainerDied","Data":"794ab5ae835235f7e7658e998c2546d0f8d0c033ee19c70e79f6e72aba7ddf02"} Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.132387 4750 scope.go:117] "RemoveContainer" containerID="e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.153614 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5be93415-f615-42a7-948f-68829841543e" (UID: "5be93415-f615-42a7-948f-68829841543e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.166678 4750 scope.go:117] "RemoveContainer" containerID="0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.174389 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.196121 4750 scope.go:117] "RemoveContainer" containerID="e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead" Mar 09 18:48:18 crc kubenswrapper[4750]: E0309 18:48:18.205899 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead\": container with ID starting with e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead not found: ID does not exist" containerID="e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.206000 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead"} err="failed to get container status \"e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead\": rpc error: code = NotFound desc = could not find container \"e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead\": container with ID starting with e41b10db8bba146b6adaa546d2d2ab47a2d94fd6c7f7a693742d25a381a99ead not found: ID does not exist" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.206062 4750 scope.go:117] "RemoveContainer" containerID="0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0" Mar 09 18:48:18 crc kubenswrapper[4750]: E0309 18:48:18.206832 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0\": container with ID starting with 0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0 not found: ID does not exist" containerID="0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.206883 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0"} err="failed to get container status \"0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0\": rpc error: code = NotFound desc = could not find container \"0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0\": container with ID starting with 0df78212cc598c0c5698393f3a24ef29ae2016928003f1542d5f2137b0e043a0 not found: ID does not exist" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.223827 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.223874 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be93415-f615-42a7-948f-68829841543e-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.466068 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fc5774769-shw5p"] Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.474761 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fc5774769-shw5p"] Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.989145 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:18 crc kubenswrapper[4750]: I0309 18:48:18.990469 4750 scope.go:117] "RemoveContainer" containerID="8a942558352aa440b404d831fbe9e6f7cef4864baa5eef35869bb3c60f919a6a" Mar 09 18:48:18 crc kubenswrapper[4750]: E0309 18:48:18.990927 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(6796c4d9-5680-4c22-9ef0-815642b22c3f)\"" pod="openstack/watcher-decision-engine-0" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" Mar 09 18:48:19 crc kubenswrapper[4750]: I0309 18:48:19.145268 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="cinder-scheduler" containerID="cri-o://b2aef219fc788d84813540d7b11774c8bf24d266e97ec717a2d9c1cee4a0e09e" gracePeriod=30 Mar 09 18:48:19 crc kubenswrapper[4750]: I0309 18:48:19.145894 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="probe" containerID="cri-o://846929547db069fe8adbbcf0a02d640976c07c15d2e9d7eb9ef85fd81f1a0aa7" gracePeriod=30 Mar 09 18:48:19 crc kubenswrapper[4750]: I0309 18:48:19.390049 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5be93415-f615-42a7-948f-68829841543e" path="/var/lib/kubelet/pods/5be93415-f615-42a7-948f-68829841543e/volumes" Mar 09 18:48:20 crc kubenswrapper[4750]: I0309 18:48:20.158049 4750 generic.go:334] "Generic (PLEG): container finished" podID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerID="846929547db069fe8adbbcf0a02d640976c07c15d2e9d7eb9ef85fd81f1a0aa7" exitCode=0 Mar 09 18:48:20 crc kubenswrapper[4750]: I0309 18:48:20.158172 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"efd2c1d8-3a05-412b-9854-1086c8002cf4","Type":"ContainerDied","Data":"846929547db069fe8adbbcf0a02d640976c07c15d2e9d7eb9ef85fd81f1a0aa7"} Mar 09 18:48:20 crc kubenswrapper[4750]: I0309 18:48:20.195367 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.172483 4750 generic.go:334] "Generic (PLEG): container finished" podID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerID="b2aef219fc788d84813540d7b11774c8bf24d266e97ec717a2d9c1cee4a0e09e" exitCode=0 Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.172706 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"efd2c1d8-3a05-412b-9854-1086c8002cf4","Type":"ContainerDied","Data":"b2aef219fc788d84813540d7b11774c8bf24d266e97ec717a2d9c1cee4a0e09e"} Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.172885 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"efd2c1d8-3a05-412b-9854-1086c8002cf4","Type":"ContainerDied","Data":"799ba6a5cdd97f2bd8441f4deb6680f8bd1ad86215dffff2bdcdc4ec6cf62b50"} Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.172904 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="799ba6a5cdd97f2bd8441f4deb6680f8bd1ad86215dffff2bdcdc4ec6cf62b50" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.218768 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.241582 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.250137 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290058 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/efd2c1d8-3a05-412b-9854-1086c8002cf4-etc-machine-id\") pod \"efd2c1d8-3a05-412b-9854-1086c8002cf4\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290109 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data-custom\") pod \"efd2c1d8-3a05-412b-9854-1086c8002cf4\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290151 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnj7z\" (UniqueName: \"kubernetes.io/projected/efd2c1d8-3a05-412b-9854-1086c8002cf4-kube-api-access-fnj7z\") pod \"efd2c1d8-3a05-412b-9854-1086c8002cf4\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290167 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efd2c1d8-3a05-412b-9854-1086c8002cf4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "efd2c1d8-3a05-412b-9854-1086c8002cf4" (UID: "efd2c1d8-3a05-412b-9854-1086c8002cf4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290360 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-combined-ca-bundle\") pod \"efd2c1d8-3a05-412b-9854-1086c8002cf4\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290432 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-scripts\") pod \"efd2c1d8-3a05-412b-9854-1086c8002cf4\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290466 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data\") pod \"efd2c1d8-3a05-412b-9854-1086c8002cf4\" (UID: \"efd2c1d8-3a05-412b-9854-1086c8002cf4\") " Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.290989 4750 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/efd2c1d8-3a05-412b-9854-1086c8002cf4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.310273 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efd2c1d8-3a05-412b-9854-1086c8002cf4-kube-api-access-fnj7z" (OuterVolumeSpecName: "kube-api-access-fnj7z") pod "efd2c1d8-3a05-412b-9854-1086c8002cf4" (UID: "efd2c1d8-3a05-412b-9854-1086c8002cf4"). InnerVolumeSpecName "kube-api-access-fnj7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.321906 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-scripts" (OuterVolumeSpecName: "scripts") pod "efd2c1d8-3a05-412b-9854-1086c8002cf4" (UID: "efd2c1d8-3a05-412b-9854-1086c8002cf4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.328732 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "efd2c1d8-3a05-412b-9854-1086c8002cf4" (UID: "efd2c1d8-3a05-412b-9854-1086c8002cf4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.371106 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efd2c1d8-3a05-412b-9854-1086c8002cf4" (UID: "efd2c1d8-3a05-412b-9854-1086c8002cf4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.399041 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.399080 4750 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.399095 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnj7z\" (UniqueName: \"kubernetes.io/projected/efd2c1d8-3a05-412b-9854-1086c8002cf4-kube-api-access-fnj7z\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.399108 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.413760 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data" (OuterVolumeSpecName: "config-data") pod "efd2c1d8-3a05-412b-9854-1086c8002cf4" (UID: "efd2c1d8-3a05-412b-9854-1086c8002cf4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:21 crc kubenswrapper[4750]: I0309 18:48:21.501063 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efd2c1d8-3a05-412b-9854-1086c8002cf4-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.189065 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.247061 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.278003 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.297960 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.318397 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.318956 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api-log" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.318981 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api-log" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.318993 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="probe" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319001 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="probe" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.319017 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="cinder-scheduler" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319025 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="cinder-scheduler" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.319041 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-api" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319049 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-api" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.319058 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319065 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.319080 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be93415-f615-42a7-948f-68829841543e" containerName="init" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319086 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be93415-f615-42a7-948f-68829841543e" containerName="init" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.319100 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d81c74f1-ee19-4573-8582-e2fdf6373472" containerName="oc" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319109 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d81c74f1-ee19-4573-8582-e2fdf6373472" containerName="oc" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.319122 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-httpd" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319131 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-httpd" Mar 09 18:48:22 crc kubenswrapper[4750]: E0309 18:48:22.319142 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be93415-f615-42a7-948f-68829841543e" containerName="dnsmasq-dns" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319151 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be93415-f615-42a7-948f-68829841543e" containerName="dnsmasq-dns" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319398 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="cinder-scheduler" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319417 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5be93415-f615-42a7-948f-68829841543e" containerName="dnsmasq-dns" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319430 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d81c74f1-ee19-4573-8582-e2fdf6373472" containerName="oc" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319441 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-httpd" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319459 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" containerName="probe" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319475 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319485 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="150ab1b1-103a-463a-9c5d-fc4231e42674" containerName="barbican-api-log" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.319497 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="72dade1d-7fb7-4118-a665-7b2fdfc56a9e" containerName="neutron-api" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.320902 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.330929 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.383568 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.432245 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgvdt\" (UniqueName: \"kubernetes.io/projected/adc68d49-e0cb-4adf-abce-56a9eae4f937-kube-api-access-jgvdt\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.432417 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.432584 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.432652 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adc68d49-e0cb-4adf-abce-56a9eae4f937-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.432676 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.432757 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-scripts\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.544698 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgvdt\" (UniqueName: \"kubernetes.io/projected/adc68d49-e0cb-4adf-abce-56a9eae4f937-kube-api-access-jgvdt\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.544790 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.544850 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.544882 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adc68d49-e0cb-4adf-abce-56a9eae4f937-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.544916 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.544955 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-scripts\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.545764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adc68d49-e0cb-4adf-abce-56a9eae4f937-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.552183 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.557111 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.559786 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-scripts\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.561754 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgvdt\" (UniqueName: \"kubernetes.io/projected/adc68d49-e0cb-4adf-abce-56a9eae4f937-kube-api-access-jgvdt\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.580304 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " pod="openstack/cinder-scheduler-0" Mar 09 18:48:22 crc kubenswrapper[4750]: I0309 18:48:22.685596 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 18:48:23 crc kubenswrapper[4750]: I0309 18:48:23.097137 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:48:23 crc kubenswrapper[4750]: I0309 18:48:23.144528 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:48:23 crc kubenswrapper[4750]: I0309 18:48:23.293191 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 18:48:23 crc kubenswrapper[4750]: I0309 18:48:23.348939 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.167:8443: connect: connection refused" Mar 09 18:48:23 crc kubenswrapper[4750]: I0309 18:48:23.395024 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efd2c1d8-3a05-412b-9854-1086c8002cf4" path="/var/lib/kubelet/pods/efd2c1d8-3a05-412b-9854-1086c8002cf4/volumes" Mar 09 18:48:24 crc kubenswrapper[4750]: I0309 18:48:24.223856 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"adc68d49-e0cb-4adf-abce-56a9eae4f937","Type":"ContainerStarted","Data":"76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2"} Mar 09 18:48:24 crc kubenswrapper[4750]: I0309 18:48:24.224088 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"adc68d49-e0cb-4adf-abce-56a9eae4f937","Type":"ContainerStarted","Data":"e62571cb9e4ef6eac1bb15cc9b927d863f8b08226d9f6052bb0a32100ef3a180"} Mar 09 18:48:24 crc kubenswrapper[4750]: I0309 18:48:24.734120 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6594b67854-xmvs4" Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.241080 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"adc68d49-e0cb-4adf-abce-56a9eae4f937","Type":"ContainerStarted","Data":"65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9"} Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.279701 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.279678469 podStartE2EDuration="3.279678469s" podCreationTimestamp="2026-03-09 18:48:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:25.274920649 +0000 UTC m=+1386.617393047" watchObservedRunningTime="2026-03-09 18:48:25.279678469 +0000 UTC m=+1386.622150877" Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.901022 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.903790 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.908243 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.908520 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bczpw" Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.908727 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 09 18:48:25 crc kubenswrapper[4750]: I0309 18:48:25.935279 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.029366 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.029598 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api-log" containerID="cri-o://4bb993426867ef727ad76e8b4ef6a9d448bb3a2b6834c1bde63f1c8a968deda4" gracePeriod=30 Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.030073 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api" containerID="cri-o://4aa97216ea9a86b28cadd4845f784fe4f9363054e5dded4e036d36e36dbc706c" gracePeriod=30 Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.038466 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkpfb\" (UniqueName: \"kubernetes.io/projected/be47c3f6-61fd-45da-b720-e1df25dd5873-kube-api-access-nkpfb\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.038679 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be47c3f6-61fd-45da-b720-e1df25dd5873-openstack-config\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.038726 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be47c3f6-61fd-45da-b720-e1df25dd5873-combined-ca-bundle\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.038771 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be47c3f6-61fd-45da-b720-e1df25dd5873-openstack-config-secret\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.136068 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.142823 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be47c3f6-61fd-45da-b720-e1df25dd5873-openstack-config\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.142988 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be47c3f6-61fd-45da-b720-e1df25dd5873-combined-ca-bundle\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.143128 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be47c3f6-61fd-45da-b720-e1df25dd5873-openstack-config-secret\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.143298 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkpfb\" (UniqueName: \"kubernetes.io/projected/be47c3f6-61fd-45da-b720-e1df25dd5873-kube-api-access-nkpfb\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.143905 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be47c3f6-61fd-45da-b720-e1df25dd5873-openstack-config\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.155238 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be47c3f6-61fd-45da-b720-e1df25dd5873-combined-ca-bundle\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.164063 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be47c3f6-61fd-45da-b720-e1df25dd5873-openstack-config-secret\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.176329 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkpfb\" (UniqueName: \"kubernetes.io/projected/be47c3f6-61fd-45da-b720-e1df25dd5873-kube-api-access-nkpfb\") pod \"openstackclient\" (UID: \"be47c3f6-61fd-45da-b720-e1df25dd5873\") " pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.231677 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.262826 4750 generic.go:334] "Generic (PLEG): container finished" podID="180bc81d-e02f-4754-93cf-e0345d11829d" containerID="4bb993426867ef727ad76e8b4ef6a9d448bb3a2b6834c1bde63f1c8a968deda4" exitCode=143 Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.264515 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"180bc81d-e02f-4754-93cf-e0345d11829d","Type":"ContainerDied","Data":"4bb993426867ef727ad76e8b4ef6a9d448bb3a2b6834c1bde63f1c8a968deda4"} Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.428804 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-64cbf4d6d8-cn7dn" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.529922 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-84fc97847d-hvzvz"] Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.530425 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-84fc97847d-hvzvz" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-log" containerID="cri-o://e2882613115715bdbbcfe0e1578ce726f508c3cef05a9f763262a8b5670bee9d" gracePeriod=30 Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.530915 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-84fc97847d-hvzvz" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-api" containerID="cri-o://1b0e63691cd35c17ed5ac5184ebd75d053c96523f994a79f33733b5aa0d34778" gracePeriod=30 Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.829871 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.929873 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.193:9322/\": read tcp 10.217.0.2:53644->10.217.0.193:9322: read: connection reset by peer" Mar 09 18:48:26 crc kubenswrapper[4750]: I0309 18:48:26.930245 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.193:9322/\": read tcp 10.217.0.2:53646->10.217.0.193:9322: read: connection reset by peer" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.294170 4750 generic.go:334] "Generic (PLEG): container finished" podID="180bc81d-e02f-4754-93cf-e0345d11829d" containerID="4aa97216ea9a86b28cadd4845f784fe4f9363054e5dded4e036d36e36dbc706c" exitCode=0 Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.295733 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"180bc81d-e02f-4754-93cf-e0345d11829d","Type":"ContainerDied","Data":"4aa97216ea9a86b28cadd4845f784fe4f9363054e5dded4e036d36e36dbc706c"} Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.298290 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"be47c3f6-61fd-45da-b720-e1df25dd5873","Type":"ContainerStarted","Data":"285082fc20c32bb920807003a19729c4b04ec4911d7e9cdc18028f2fa0f7a7d7"} Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.306780 4750 generic.go:334] "Generic (PLEG): container finished" podID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerID="e2882613115715bdbbcfe0e1578ce726f508c3cef05a9f763262a8b5670bee9d" exitCode=143 Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.310085 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84fc97847d-hvzvz" event={"ID":"edd199bd-36e4-4f32-a4cf-237d7a88011a","Type":"ContainerDied","Data":"e2882613115715bdbbcfe0e1578ce726f508c3cef05a9f763262a8b5670bee9d"} Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.435120 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.586999 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-config-data\") pod \"180bc81d-e02f-4754-93cf-e0345d11829d\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.587136 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-combined-ca-bundle\") pod \"180bc81d-e02f-4754-93cf-e0345d11829d\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.587252 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/180bc81d-e02f-4754-93cf-e0345d11829d-logs\") pod \"180bc81d-e02f-4754-93cf-e0345d11829d\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.587282 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-custom-prometheus-ca\") pod \"180bc81d-e02f-4754-93cf-e0345d11829d\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.587323 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvb4b\" (UniqueName: \"kubernetes.io/projected/180bc81d-e02f-4754-93cf-e0345d11829d-kube-api-access-vvb4b\") pod \"180bc81d-e02f-4754-93cf-e0345d11829d\" (UID: \"180bc81d-e02f-4754-93cf-e0345d11829d\") " Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.589281 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/180bc81d-e02f-4754-93cf-e0345d11829d-logs" (OuterVolumeSpecName: "logs") pod "180bc81d-e02f-4754-93cf-e0345d11829d" (UID: "180bc81d-e02f-4754-93cf-e0345d11829d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.628383 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180bc81d-e02f-4754-93cf-e0345d11829d-kube-api-access-vvb4b" (OuterVolumeSpecName: "kube-api-access-vvb4b") pod "180bc81d-e02f-4754-93cf-e0345d11829d" (UID: "180bc81d-e02f-4754-93cf-e0345d11829d"). InnerVolumeSpecName "kube-api-access-vvb4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.661696 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "180bc81d-e02f-4754-93cf-e0345d11829d" (UID: "180bc81d-e02f-4754-93cf-e0345d11829d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.693928 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.696070 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.696098 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/180bc81d-e02f-4754-93cf-e0345d11829d-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.696109 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvb4b\" (UniqueName: \"kubernetes.io/projected/180bc81d-e02f-4754-93cf-e0345d11829d-kube-api-access-vvb4b\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.748524 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "180bc81d-e02f-4754-93cf-e0345d11829d" (UID: "180bc81d-e02f-4754-93cf-e0345d11829d"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.778399 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-config-data" (OuterVolumeSpecName: "config-data") pod "180bc81d-e02f-4754-93cf-e0345d11829d" (UID: "180bc81d-e02f-4754-93cf-e0345d11829d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.800057 4750 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:27 crc kubenswrapper[4750]: I0309 18:48:27.800096 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180bc81d-e02f-4754-93cf-e0345d11829d-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.328022 4750 generic.go:334] "Generic (PLEG): container finished" podID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerID="1b0e63691cd35c17ed5ac5184ebd75d053c96523f994a79f33733b5aa0d34778" exitCode=0 Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.328407 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84fc97847d-hvzvz" event={"ID":"edd199bd-36e4-4f32-a4cf-237d7a88011a","Type":"ContainerDied","Data":"1b0e63691cd35c17ed5ac5184ebd75d053c96523f994a79f33733b5aa0d34778"} Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.333583 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"180bc81d-e02f-4754-93cf-e0345d11829d","Type":"ContainerDied","Data":"3e897e6ae1c06521196e8de8fae3d284a1d7437623b6519e2e1faac6a014ee52"} Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.333658 4750 scope.go:117] "RemoveContainer" containerID="4aa97216ea9a86b28cadd4845f784fe4f9363054e5dded4e036d36e36dbc706c" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.333800 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.391575 4750 scope.go:117] "RemoveContainer" containerID="4bb993426867ef727ad76e8b4ef6a9d448bb3a2b6834c1bde63f1c8a968deda4" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.462568 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.469614 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.483270 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:28 crc kubenswrapper[4750]: E0309 18:48:28.485749 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api-log" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.485772 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api-log" Mar 09 18:48:28 crc kubenswrapper[4750]: E0309 18:48:28.485790 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.485797 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.486038 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api-log" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.486051 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" containerName="watcher-api" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.489682 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.490438 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.493141 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.494121 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.494354 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.627960 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.628032 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.628060 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-public-tls-certs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.628307 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-config-data\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.628394 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8060c58-bb40-45af-86d4-310cafdcfde6-logs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.628611 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6jxq\" (UniqueName: \"kubernetes.io/projected/e8060c58-bb40-45af-86d4-310cafdcfde6-kube-api-access-x6jxq\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.628698 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730329 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8060c58-bb40-45af-86d4-310cafdcfde6-logs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730434 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6jxq\" (UniqueName: \"kubernetes.io/projected/e8060c58-bb40-45af-86d4-310cafdcfde6-kube-api-access-x6jxq\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730462 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730526 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730583 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730602 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-public-tls-certs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730657 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-config-data\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.730796 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8060c58-bb40-45af-86d4-310cafdcfde6-logs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.737783 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.738012 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.738349 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-public-tls-certs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.740246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-config-data\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.744590 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.745156 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.754347 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6jxq\" (UniqueName: \"kubernetes.io/projected/e8060c58-bb40-45af-86d4-310cafdcfde6-kube-api-access-x6jxq\") pod \"watcher-api-0\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.827793 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.940234 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-combined-ca-bundle\") pod \"edd199bd-36e4-4f32-a4cf-237d7a88011a\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.940428 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlbqn\" (UniqueName: \"kubernetes.io/projected/edd199bd-36e4-4f32-a4cf-237d7a88011a-kube-api-access-vlbqn\") pod \"edd199bd-36e4-4f32-a4cf-237d7a88011a\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.940453 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-internal-tls-certs\") pod \"edd199bd-36e4-4f32-a4cf-237d7a88011a\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.943799 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-scripts\") pod \"edd199bd-36e4-4f32-a4cf-237d7a88011a\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.943839 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd199bd-36e4-4f32-a4cf-237d7a88011a-logs\") pod \"edd199bd-36e4-4f32-a4cf-237d7a88011a\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.943910 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-config-data\") pod \"edd199bd-36e4-4f32-a4cf-237d7a88011a\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.944053 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-public-tls-certs\") pod \"edd199bd-36e4-4f32-a4cf-237d7a88011a\" (UID: \"edd199bd-36e4-4f32-a4cf-237d7a88011a\") " Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.949204 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-scripts" (OuterVolumeSpecName: "scripts") pod "edd199bd-36e4-4f32-a4cf-237d7a88011a" (UID: "edd199bd-36e4-4f32-a4cf-237d7a88011a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.955650 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edd199bd-36e4-4f32-a4cf-237d7a88011a-logs" (OuterVolumeSpecName: "logs") pod "edd199bd-36e4-4f32-a4cf-237d7a88011a" (UID: "edd199bd-36e4-4f32-a4cf-237d7a88011a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.961279 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edd199bd-36e4-4f32-a4cf-237d7a88011a-kube-api-access-vlbqn" (OuterVolumeSpecName: "kube-api-access-vlbqn") pod "edd199bd-36e4-4f32-a4cf-237d7a88011a" (UID: "edd199bd-36e4-4f32-a4cf-237d7a88011a"). InnerVolumeSpecName "kube-api-access-vlbqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.990855 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.990928 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:28 crc kubenswrapper[4750]: I0309 18:48:28.991831 4750 scope.go:117] "RemoveContainer" containerID="8a942558352aa440b404d831fbe9e6f7cef4864baa5eef35869bb3c60f919a6a" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.037744 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-config-data" (OuterVolumeSpecName: "config-data") pod "edd199bd-36e4-4f32-a4cf-237d7a88011a" (UID: "edd199bd-36e4-4f32-a4cf-237d7a88011a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.050870 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlbqn\" (UniqueName: \"kubernetes.io/projected/edd199bd-36e4-4f32-a4cf-237d7a88011a-kube-api-access-vlbqn\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.050900 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.050911 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd199bd-36e4-4f32-a4cf-237d7a88011a-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.050919 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.077770 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edd199bd-36e4-4f32-a4cf-237d7a88011a" (UID: "edd199bd-36e4-4f32-a4cf-237d7a88011a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.142729 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "edd199bd-36e4-4f32-a4cf-237d7a88011a" (UID: "edd199bd-36e4-4f32-a4cf-237d7a88011a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.154349 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.154495 4750 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.155422 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "edd199bd-36e4-4f32-a4cf-237d7a88011a" (UID: "edd199bd-36e4-4f32-a4cf-237d7a88011a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.256112 4750 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd199bd-36e4-4f32-a4cf-237d7a88011a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.320341 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.350528 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerStarted","Data":"eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f"} Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.358848 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84fc97847d-hvzvz" event={"ID":"edd199bd-36e4-4f32-a4cf-237d7a88011a","Type":"ContainerDied","Data":"dd07a98c706b499332baba6944b47826b2b92713f43d2ade90b9ba0a549bcd34"} Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.358915 4750 scope.go:117] "RemoveContainer" containerID="1b0e63691cd35c17ed5ac5184ebd75d053c96523f994a79f33733b5aa0d34778" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.359087 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84fc97847d-hvzvz" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.415563 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="180bc81d-e02f-4754-93cf-e0345d11829d" path="/var/lib/kubelet/pods/180bc81d-e02f-4754-93cf-e0345d11829d/volumes" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.459883 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-84fc97847d-hvzvz"] Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.465671 4750 scope.go:117] "RemoveContainer" containerID="e2882613115715bdbbcfe0e1578ce726f508c3cef05a9f763262a8b5670bee9d" Mar 09 18:48:29 crc kubenswrapper[4750]: I0309 18:48:29.469570 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-84fc97847d-hvzvz"] Mar 09 18:48:30 crc kubenswrapper[4750]: I0309 18:48:30.373437 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e8060c58-bb40-45af-86d4-310cafdcfde6","Type":"ContainerStarted","Data":"31ac876a95bf09d01ad8b09379fa6625140fab0337498dd49db09774a709a665"} Mar 09 18:48:30 crc kubenswrapper[4750]: I0309 18:48:30.373866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e8060c58-bb40-45af-86d4-310cafdcfde6","Type":"ContainerStarted","Data":"81fa205da7414242348162b76897b36b68057304538aeb66e5fce95f2da52539"} Mar 09 18:48:30 crc kubenswrapper[4750]: I0309 18:48:30.373889 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:48:30 crc kubenswrapper[4750]: I0309 18:48:30.373947 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e8060c58-bb40-45af-86d4-310cafdcfde6","Type":"ContainerStarted","Data":"16c80cf1aa66d72a714d62df4fbecd17f022e8e61530e69c0086a3a396f3a207"} Mar 09 18:48:30 crc kubenswrapper[4750]: I0309 18:48:30.397068 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=2.397050929 podStartE2EDuration="2.397050929s" podCreationTimestamp="2026-03-09 18:48:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:30.392380842 +0000 UTC m=+1391.734853240" watchObservedRunningTime="2026-03-09 18:48:30.397050929 +0000 UTC m=+1391.739523327" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.387056 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" path="/var/lib/kubelet/pods/edd199bd-36e4-4f32-a4cf-237d7a88011a/volumes" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.523198 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-54cf6684f5-bz5sx"] Mar 09 18:48:31 crc kubenswrapper[4750]: E0309 18:48:31.523625 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-log" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.523659 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-log" Mar 09 18:48:31 crc kubenswrapper[4750]: E0309 18:48:31.523670 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-api" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.523676 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-api" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.523877 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-api" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.523901 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd199bd-36e4-4f32-a4cf-237d7a88011a" containerName="placement-log" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.524961 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.538160 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.538361 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.538658 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.552879 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-54cf6684f5-bz5sx"] Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720025 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-run-httpd\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720089 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-log-httpd\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720245 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-internal-tls-certs\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720300 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thb42\" (UniqueName: \"kubernetes.io/projected/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-kube-api-access-thb42\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720351 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-public-tls-certs\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720390 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-config-data\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720420 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-etc-swift\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.720473 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-combined-ca-bundle\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822213 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-etc-swift\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822281 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-combined-ca-bundle\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822320 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-run-httpd\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-log-httpd\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822444 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-internal-tls-certs\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822461 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thb42\" (UniqueName: \"kubernetes.io/projected/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-kube-api-access-thb42\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822492 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-public-tls-certs\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822520 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-config-data\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822885 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-run-httpd\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.822951 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-log-httpd\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.827725 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-public-tls-certs\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.829192 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-config-data\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.834460 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-combined-ca-bundle\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.841345 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thb42\" (UniqueName: \"kubernetes.io/projected/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-kube-api-access-thb42\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.847095 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-etc-swift\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.851499 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96-internal-tls-certs\") pod \"swift-proxy-54cf6684f5-bz5sx\" (UID: \"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96\") " pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:31 crc kubenswrapper[4750]: I0309 18:48:31.861091 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:32 crc kubenswrapper[4750]: I0309 18:48:32.496051 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-54cf6684f5-bz5sx"] Mar 09 18:48:32 crc kubenswrapper[4750]: I0309 18:48:32.880530 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.255082 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.349347 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5ddc4b4b8d-f5q7t" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.167:8443: connect: connection refused" Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.417116 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-54cf6684f5-bz5sx" event={"ID":"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96","Type":"ContainerStarted","Data":"2f3487159a4b63eaed37887c481e0ecb7b535b65b3583fe490e746fa819c0300"} Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.417186 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-54cf6684f5-bz5sx" event={"ID":"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96","Type":"ContainerStarted","Data":"9bb84e734a4d90d23b36447d2eb8e74ccda23d508cb100d659659c1e68f1ed39"} Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.417200 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-54cf6684f5-bz5sx" event={"ID":"a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96","Type":"ContainerStarted","Data":"ec68850ab5d83b83175ae829ff9606c30e97463bcf4af6bb259ed6c022419134"} Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.417315 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.448260 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-54cf6684f5-bz5sx" podStartSLOduration=2.448238804 podStartE2EDuration="2.448238804s" podCreationTimestamp="2026-03-09 18:48:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:33.435993741 +0000 UTC m=+1394.778466139" watchObservedRunningTime="2026-03-09 18:48:33.448238804 +0000 UTC m=+1394.790711202" Mar 09 18:48:33 crc kubenswrapper[4750]: I0309 18:48:33.829016 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.258790 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.313921 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-qn2pj"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.315272 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.322137 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-qn2pj"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.427420 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-j4pfx"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.428888 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.447590 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-08df-account-create-update-9t8fq"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.449187 4750 generic.go:334] "Generic (PLEG): container finished" podID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" exitCode=1 Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.450759 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerDied","Data":"eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f"} Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.450812 4750 scope.go:117] "RemoveContainer" containerID="8a942558352aa440b404d831fbe9e6f7cef4864baa5eef35869bb3c60f919a6a" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.450949 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.451886 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.452847 4750 scope.go:117] "RemoveContainer" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" Mar 09 18:48:34 crc kubenswrapper[4750]: E0309 18:48:34.453101 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(6796c4d9-5680-4c22-9ef0-815642b22c3f)\"" pod="openstack/watcher-decision-engine-0" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.461400 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.492366 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3837950-d435-4f57-9bde-3c5172cba1d8-operator-scripts\") pod \"nova-api-db-create-qn2pj\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.492518 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqm6j\" (UniqueName: \"kubernetes.io/projected/e3837950-d435-4f57-9bde-3c5172cba1d8-kube-api-access-jqm6j\") pod \"nova-api-db-create-qn2pj\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.494699 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-j4pfx"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.506135 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-08df-account-create-update-9t8fq"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.542831 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dk8lh"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.544123 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.553830 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dk8lh"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.594684 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9rb9\" (UniqueName: \"kubernetes.io/projected/081630cf-500b-49f1-992d-233ed78ff1bc-kube-api-access-c9rb9\") pod \"nova-api-08df-account-create-update-9t8fq\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.594764 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3837950-d435-4f57-9bde-3c5172cba1d8-operator-scripts\") pod \"nova-api-db-create-qn2pj\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.594806 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081630cf-500b-49f1-992d-233ed78ff1bc-operator-scripts\") pod \"nova-api-08df-account-create-update-9t8fq\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.594862 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-operator-scripts\") pod \"nova-cell0-db-create-j4pfx\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.594943 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqm6j\" (UniqueName: \"kubernetes.io/projected/e3837950-d435-4f57-9bde-3c5172cba1d8-kube-api-access-jqm6j\") pod \"nova-api-db-create-qn2pj\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.594972 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2vq2\" (UniqueName: \"kubernetes.io/projected/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-kube-api-access-h2vq2\") pod \"nova-cell0-db-create-j4pfx\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.596816 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3837950-d435-4f57-9bde-3c5172cba1d8-operator-scripts\") pod \"nova-api-db-create-qn2pj\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.610649 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b786-account-create-update-dr9fd"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.611859 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.616045 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.620959 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqm6j\" (UniqueName: \"kubernetes.io/projected/e3837950-d435-4f57-9bde-3c5172cba1d8-kube-api-access-jqm6j\") pod \"nova-api-db-create-qn2pj\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.654658 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.661511 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b786-account-create-update-dr9fd"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697267 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwjvx\" (UniqueName: \"kubernetes.io/projected/f9746db5-1458-4326-ba1f-5b3abbecce78-kube-api-access-mwjvx\") pod \"nova-cell0-b786-account-create-update-dr9fd\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b2776f-8175-48d4-8c86-9269df88aa64-operator-scripts\") pod \"nova-cell1-db-create-dk8lh\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697600 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9rb9\" (UniqueName: \"kubernetes.io/projected/081630cf-500b-49f1-992d-233ed78ff1bc-kube-api-access-c9rb9\") pod \"nova-api-08df-account-create-update-9t8fq\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697664 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081630cf-500b-49f1-992d-233ed78ff1bc-operator-scripts\") pod \"nova-api-08df-account-create-update-9t8fq\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697695 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-776r8\" (UniqueName: \"kubernetes.io/projected/e3b2776f-8175-48d4-8c86-9269df88aa64-kube-api-access-776r8\") pod \"nova-cell1-db-create-dk8lh\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697750 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-operator-scripts\") pod \"nova-cell0-db-create-j4pfx\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697841 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9746db5-1458-4326-ba1f-5b3abbecce78-operator-scripts\") pod \"nova-cell0-b786-account-create-update-dr9fd\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.697868 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2vq2\" (UniqueName: \"kubernetes.io/projected/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-kube-api-access-h2vq2\") pod \"nova-cell0-db-create-j4pfx\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.698971 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081630cf-500b-49f1-992d-233ed78ff1bc-operator-scripts\") pod \"nova-api-08df-account-create-update-9t8fq\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.699344 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-operator-scripts\") pod \"nova-cell0-db-create-j4pfx\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.724274 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9rb9\" (UniqueName: \"kubernetes.io/projected/081630cf-500b-49f1-992d-233ed78ff1bc-kube-api-access-c9rb9\") pod \"nova-api-08df-account-create-update-9t8fq\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.737199 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2vq2\" (UniqueName: \"kubernetes.io/projected/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-kube-api-access-h2vq2\") pod \"nova-cell0-db-create-j4pfx\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.766082 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.801426 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.802531 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9746db5-1458-4326-ba1f-5b3abbecce78-operator-scripts\") pod \"nova-cell0-b786-account-create-update-dr9fd\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.802734 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwjvx\" (UniqueName: \"kubernetes.io/projected/f9746db5-1458-4326-ba1f-5b3abbecce78-kube-api-access-mwjvx\") pod \"nova-cell0-b786-account-create-update-dr9fd\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.802823 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b2776f-8175-48d4-8c86-9269df88aa64-operator-scripts\") pod \"nova-cell1-db-create-dk8lh\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.802934 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-776r8\" (UniqueName: \"kubernetes.io/projected/e3b2776f-8175-48d4-8c86-9269df88aa64-kube-api-access-776r8\") pod \"nova-cell1-db-create-dk8lh\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.804059 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9746db5-1458-4326-ba1f-5b3abbecce78-operator-scripts\") pod \"nova-cell0-b786-account-create-update-dr9fd\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.804244 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b2776f-8175-48d4-8c86-9269df88aa64-operator-scripts\") pod \"nova-cell1-db-create-dk8lh\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.833279 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwjvx\" (UniqueName: \"kubernetes.io/projected/f9746db5-1458-4326-ba1f-5b3abbecce78-kube-api-access-mwjvx\") pod \"nova-cell0-b786-account-create-update-dr9fd\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.850415 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5f67-account-create-update-l2t7v"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.851848 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.860297 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5f67-account-create-update-l2t7v"] Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.861023 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.862912 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-776r8\" (UniqueName: \"kubernetes.io/projected/e3b2776f-8175-48d4-8c86-9269df88aa64-kube-api-access-776r8\") pod \"nova-cell1-db-create-dk8lh\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.868191 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:34 crc kubenswrapper[4750]: I0309 18:48:34.990116 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:35 crc kubenswrapper[4750]: I0309 18:48:35.009414 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bghp4\" (UniqueName: \"kubernetes.io/projected/55185786-16ac-45a0-90b2-e44d0e833c23-kube-api-access-bghp4\") pod \"nova-cell1-5f67-account-create-update-l2t7v\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:35 crc kubenswrapper[4750]: I0309 18:48:35.009585 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55185786-16ac-45a0-90b2-e44d0e833c23-operator-scripts\") pod \"nova-cell1-5f67-account-create-update-l2t7v\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:35 crc kubenswrapper[4750]: I0309 18:48:35.111929 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bghp4\" (UniqueName: \"kubernetes.io/projected/55185786-16ac-45a0-90b2-e44d0e833c23-kube-api-access-bghp4\") pod \"nova-cell1-5f67-account-create-update-l2t7v\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:35 crc kubenswrapper[4750]: I0309 18:48:35.112074 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55185786-16ac-45a0-90b2-e44d0e833c23-operator-scripts\") pod \"nova-cell1-5f67-account-create-update-l2t7v\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:35 crc kubenswrapper[4750]: I0309 18:48:35.112873 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55185786-16ac-45a0-90b2-e44d0e833c23-operator-scripts\") pod \"nova-cell1-5f67-account-create-update-l2t7v\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:35 crc kubenswrapper[4750]: I0309 18:48:35.127441 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bghp4\" (UniqueName: \"kubernetes.io/projected/55185786-16ac-45a0-90b2-e44d0e833c23-kube-api-access-bghp4\") pod \"nova-cell1-5f67-account-create-update-l2t7v\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:35 crc kubenswrapper[4750]: I0309 18:48:35.239080 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.441283 4750 scope.go:117] "RemoveContainer" containerID="475b8bc49d7bc443a4a1a95a0b3ede890a53eea2c72957ed06a64305689d8abc" Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.501269 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerID="47db330f58e384d75e7b5a4e8eabdd01fc7f8f80577c7d541f52eb618d160dd7" exitCode=137 Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.501314 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerDied","Data":"47db330f58e384d75e7b5a4e8eabdd01fc7f8f80577c7d541f52eb618d160dd7"} Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.676964 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cd6d4c84f-zrhsg" Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.764652 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d6db5794d-f9pxj"] Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.764912 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d6db5794d-f9pxj" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-api" containerID="cri-o://5ad8d0972171dc5644b0c875da08f240816d847855e0e2602a5e8575930a3e3e" gracePeriod=30 Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.765048 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d6db5794d-f9pxj" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-httpd" containerID="cri-o://f9d8d985d34814bb2928cc6886169310c947dcea0cf7abd0d8398aa3e3a01f7c" gracePeriod=30 Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.828968 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.845105 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.989474 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.989555 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:38 crc kubenswrapper[4750]: I0309 18:48:38.990689 4750 scope.go:117] "RemoveContainer" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" Mar 09 18:48:38 crc kubenswrapper[4750]: E0309 18:48:38.991126 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(6796c4d9-5680-4c22-9ef0-815642b22c3f)\"" pod="openstack/watcher-decision-engine-0" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" Mar 09 18:48:39 crc kubenswrapper[4750]: I0309 18:48:39.511663 4750 generic.go:334] "Generic (PLEG): container finished" podID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerID="f9d8d985d34814bb2928cc6886169310c947dcea0cf7abd0d8398aa3e3a01f7c" exitCode=0 Mar 09 18:48:39 crc kubenswrapper[4750]: I0309 18:48:39.511860 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6db5794d-f9pxj" event={"ID":"9d5b4b2b-6058-44ac-b24b-e5ef2423f830","Type":"ContainerDied","Data":"f9d8d985d34814bb2928cc6886169310c947dcea0cf7abd0d8398aa3e3a01f7c"} Mar 09 18:48:39 crc kubenswrapper[4750]: I0309 18:48:39.525021 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Mar 09 18:48:40 crc kubenswrapper[4750]: W0309 18:48:40.248264 4750 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6796c4d9_5680_4c22_9ef0_815642b22c3f.slice/crio-conmon-eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6796c4d9_5680_4c22_9ef0_815642b22c3f.slice/crio-conmon-eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f.scope: no such file or directory Mar 09 18:48:40 crc kubenswrapper[4750]: W0309 18:48:40.248776 4750 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6796c4d9_5680_4c22_9ef0_815642b22c3f.slice/crio-eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6796c4d9_5680_4c22_9ef0_815642b22c3f.slice/crio-eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f.scope: no such file or directory Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.297794 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.437151 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-log-httpd\") pod \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.437211 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-scripts\") pod \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.437375 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-sg-core-conf-yaml\") pod \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.437410 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64gbr\" (UniqueName: \"kubernetes.io/projected/4d6c59a9-358a-4f97-8c95-f1a23f3854be-kube-api-access-64gbr\") pod \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.437511 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-combined-ca-bundle\") pod \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.437743 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-run-httpd\") pod \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.437803 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-config-data\") pod \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\" (UID: \"4d6c59a9-358a-4f97-8c95-f1a23f3854be\") " Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.442652 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4d6c59a9-358a-4f97-8c95-f1a23f3854be" (UID: "4d6c59a9-358a-4f97-8c95-f1a23f3854be"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.444989 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4d6c59a9-358a-4f97-8c95-f1a23f3854be" (UID: "4d6c59a9-358a-4f97-8c95-f1a23f3854be"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.459327 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-scripts" (OuterVolumeSpecName: "scripts") pod "4d6c59a9-358a-4f97-8c95-f1a23f3854be" (UID: "4d6c59a9-358a-4f97-8c95-f1a23f3854be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.461911 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d6c59a9-358a-4f97-8c95-f1a23f3854be-kube-api-access-64gbr" (OuterVolumeSpecName: "kube-api-access-64gbr") pod "4d6c59a9-358a-4f97-8c95-f1a23f3854be" (UID: "4d6c59a9-358a-4f97-8c95-f1a23f3854be"). InnerVolumeSpecName "kube-api-access-64gbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.490273 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-08df-account-create-update-9t8fq"] Mar 09 18:48:40 crc kubenswrapper[4750]: E0309 18:48:40.491982 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d5b4b2b_6058_44ac_b24b_e5ef2423f830.slice/crio-f9d8d985d34814bb2928cc6886169310c947dcea0cf7abd0d8398aa3e3a01f7c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef45ec45_1ef2_46e6_bf33_0c092885679f.slice/crio-conmon-7c882d7feb763fa768bb968ae79a038f49e22c03454de50f1ddfe39b393e082b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d5b4b2b_6058_44ac_b24b_e5ef2423f830.slice/crio-conmon-f9d8d985d34814bb2928cc6886169310c947dcea0cf7abd0d8398aa3e3a01f7c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef45ec45_1ef2_46e6_bf33_0c092885679f.slice/crio-7c882d7feb763fa768bb968ae79a038f49e22c03454de50f1ddfe39b393e082b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d6c59a9_358a_4f97_8c95_f1a23f3854be.slice/crio-conmon-47db330f58e384d75e7b5a4e8eabdd01fc7f8f80577c7d541f52eb618d160dd7.scope\": RecentStats: unable to find data in memory cache]" Mar 09 18:48:40 crc kubenswrapper[4750]: W0309 18:48:40.498301 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081630cf_500b_49f1_992d_233ed78ff1bc.slice/crio-3d3c7199459b04c1ac5ae9e1a2917ca8ace837a6c02349d3f116adb4177bf53f WatchSource:0}: Error finding container 3d3c7199459b04c1ac5ae9e1a2917ca8ace837a6c02349d3f116adb4177bf53f: Status 404 returned error can't find the container with id 3d3c7199459b04c1ac5ae9e1a2917ca8ace837a6c02349d3f116adb4177bf53f Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.501768 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5f67-account-create-update-l2t7v"] Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.520562 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4d6c59a9-358a-4f97-8c95-f1a23f3854be" (UID: "4d6c59a9-358a-4f97-8c95-f1a23f3854be"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.529783 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d6c59a9-358a-4f97-8c95-f1a23f3854be" (UID: "4d6c59a9-358a-4f97-8c95-f1a23f3854be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.535646 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d6c59a9-358a-4f97-8c95-f1a23f3854be","Type":"ContainerDied","Data":"fc29b5100f86212c229e7035c2d0bc971b1f509ff517b56860b5e2f0a1900351"} Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.535706 4750 scope.go:117] "RemoveContainer" containerID="47db330f58e384d75e7b5a4e8eabdd01fc7f8f80577c7d541f52eb618d160dd7" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.535749 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.547163 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-config-data" (OuterVolumeSpecName: "config-data") pod "4d6c59a9-358a-4f97-8c95-f1a23f3854be" (UID: "4d6c59a9-358a-4f97-8c95-f1a23f3854be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.548204 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-08df-account-create-update-9t8fq" event={"ID":"081630cf-500b-49f1-992d-233ed78ff1bc","Type":"ContainerStarted","Data":"3d3c7199459b04c1ac5ae9e1a2917ca8ace837a6c02349d3f116adb4177bf53f"} Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.563328 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.565775 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64gbr\" (UniqueName: \"kubernetes.io/projected/4d6c59a9-358a-4f97-8c95-f1a23f3854be-kube-api-access-64gbr\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.565875 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.565933 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.565987 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.566047 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d6c59a9-358a-4f97-8c95-f1a23f3854be-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.566099 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d6c59a9-358a-4f97-8c95-f1a23f3854be-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.568228 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"be47c3f6-61fd-45da-b720-e1df25dd5873","Type":"ContainerStarted","Data":"5154c46727c1c08fd158dfff85e9545f2a9d8c0ba6fcfb216012eb6570650f98"} Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.571067 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" event={"ID":"55185786-16ac-45a0-90b2-e44d0e833c23","Type":"ContainerStarted","Data":"bad2db5dc077a6a7a701138b560c6a587839aab4ebbe118efe29004531375f67"} Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.588619 4750 generic.go:334] "Generic (PLEG): container finished" podID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerID="7c882d7feb763fa768bb968ae79a038f49e22c03454de50f1ddfe39b393e082b" exitCode=137 Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.589673 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddc4b4b8d-f5q7t" event={"ID":"ef45ec45-1ef2-46e6-bf33-0c092885679f","Type":"ContainerDied","Data":"7c882d7feb763fa768bb968ae79a038f49e22c03454de50f1ddfe39b393e082b"} Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.598804 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.483222986 podStartE2EDuration="15.598777992s" podCreationTimestamp="2026-03-09 18:48:25 +0000 UTC" firstStartedPulling="2026-03-09 18:48:26.837846432 +0000 UTC m=+1388.180318830" lastFinishedPulling="2026-03-09 18:48:39.953401438 +0000 UTC m=+1401.295873836" observedRunningTime="2026-03-09 18:48:40.590828886 +0000 UTC m=+1401.933301284" watchObservedRunningTime="2026-03-09 18:48:40.598777992 +0000 UTC m=+1401.941250390" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.609160 4750 scope.go:117] "RemoveContainer" containerID="5e6269255415c111f968e23ad15921e57205996484e746f79d35fde5ab0b990e" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.696248 4750 scope.go:117] "RemoveContainer" containerID="6e6034f7d85033a58557e0e8af1f961c2ebd2a12c6f963ee1aa5060d9ca53b8e" Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.714325 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-j4pfx"] Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.880285 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b786-account-create-update-dr9fd"] Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.889406 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dk8lh"] Mar 09 18:48:40 crc kubenswrapper[4750]: W0309 18:48:40.899728 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9746db5_1458_4326_ba1f_5b3abbecce78.slice/crio-c61ed5ff9dd7b38d8f5ca793e30d40ad725e369a899429ae883177f25a9b634b WatchSource:0}: Error finding container c61ed5ff9dd7b38d8f5ca793e30d40ad725e369a899429ae883177f25a9b634b: Status 404 returned error can't find the container with id c61ed5ff9dd7b38d8f5ca793e30d40ad725e369a899429ae883177f25a9b634b Mar 09 18:48:40 crc kubenswrapper[4750]: I0309 18:48:40.907127 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-qn2pj"] Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.022582 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.105711 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.121695 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142094 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:41 crc kubenswrapper[4750]: E0309 18:48:41.142539 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="ceilometer-notification-agent" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142560 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="ceilometer-notification-agent" Mar 09 18:48:41 crc kubenswrapper[4750]: E0309 18:48:41.142572 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142579 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" Mar 09 18:48:41 crc kubenswrapper[4750]: E0309 18:48:41.142595 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon-log" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142601 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon-log" Mar 09 18:48:41 crc kubenswrapper[4750]: E0309 18:48:41.142689 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="sg-core" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142698 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="sg-core" Mar 09 18:48:41 crc kubenswrapper[4750]: E0309 18:48:41.142708 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="proxy-httpd" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142714 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="proxy-httpd" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142911 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="sg-core" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142938 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="ceilometer-notification-agent" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142950 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142960 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" containerName="proxy-httpd" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.142974 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" containerName="horizon-log" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.155155 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.165022 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.167913 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.200507 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-combined-ca-bundle\") pod \"ef45ec45-1ef2-46e6-bf33-0c092885679f\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.200562 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-config-data\") pod \"ef45ec45-1ef2-46e6-bf33-0c092885679f\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.200596 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-secret-key\") pod \"ef45ec45-1ef2-46e6-bf33-0c092885679f\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.200669 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-scripts\") pod \"ef45ec45-1ef2-46e6-bf33-0c092885679f\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.200805 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qs7d\" (UniqueName: \"kubernetes.io/projected/ef45ec45-1ef2-46e6-bf33-0c092885679f-kube-api-access-6qs7d\") pod \"ef45ec45-1ef2-46e6-bf33-0c092885679f\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.200881 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-tls-certs\") pod \"ef45ec45-1ef2-46e6-bf33-0c092885679f\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.200997 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef45ec45-1ef2-46e6-bf33-0c092885679f-logs\") pod \"ef45ec45-1ef2-46e6-bf33-0c092885679f\" (UID: \"ef45ec45-1ef2-46e6-bf33-0c092885679f\") " Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.202901 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef45ec45-1ef2-46e6-bf33-0c092885679f-logs" (OuterVolumeSpecName: "logs") pod "ef45ec45-1ef2-46e6-bf33-0c092885679f" (UID: "ef45ec45-1ef2-46e6-bf33-0c092885679f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.214312 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.239050 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef45ec45-1ef2-46e6-bf33-0c092885679f-kube-api-access-6qs7d" (OuterVolumeSpecName: "kube-api-access-6qs7d") pod "ef45ec45-1ef2-46e6-bf33-0c092885679f" (UID: "ef45ec45-1ef2-46e6-bf33-0c092885679f"). InnerVolumeSpecName "kube-api-access-6qs7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.240418 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ef45ec45-1ef2-46e6-bf33-0c092885679f" (UID: "ef45ec45-1ef2-46e6-bf33-0c092885679f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304177 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-config-data\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304267 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-scripts\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304286 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-log-httpd\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304338 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-run-httpd\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304375 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfksc\" (UniqueName: \"kubernetes.io/projected/85fbfb01-ec16-4977-9f54-16b0916b93ca-kube-api-access-xfksc\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304397 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304413 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304490 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef45ec45-1ef2-46e6-bf33-0c092885679f-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304500 4750 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.304510 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qs7d\" (UniqueName: \"kubernetes.io/projected/ef45ec45-1ef2-46e6-bf33-0c092885679f-kube-api-access-6qs7d\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.336340 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-config-data" (OuterVolumeSpecName: "config-data") pod "ef45ec45-1ef2-46e6-bf33-0c092885679f" (UID: "ef45ec45-1ef2-46e6-bf33-0c092885679f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.348911 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef45ec45-1ef2-46e6-bf33-0c092885679f" (UID: "ef45ec45-1ef2-46e6-bf33-0c092885679f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.385171 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-scripts" (OuterVolumeSpecName: "scripts") pod "ef45ec45-1ef2-46e6-bf33-0c092885679f" (UID: "ef45ec45-1ef2-46e6-bf33-0c092885679f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.395978 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d6c59a9-358a-4f97-8c95-f1a23f3854be" path="/var/lib/kubelet/pods/4d6c59a9-358a-4f97-8c95-f1a23f3854be/volumes" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.408490 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-config-data\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.408585 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-scripts\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.408606 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-log-httpd\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.410095 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-run-httpd\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.410174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfksc\" (UniqueName: \"kubernetes.io/projected/85fbfb01-ec16-4977-9f54-16b0916b93ca-kube-api-access-xfksc\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.410250 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.411157 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-log-httpd\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.412154 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.412397 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.412414 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.412423 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef45ec45-1ef2-46e6-bf33-0c092885679f-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.414015 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-run-httpd\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.417245 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-config-data\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.419167 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-scripts\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.420776 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.423832 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.428700 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "ef45ec45-1ef2-46e6-bf33-0c092885679f" (UID: "ef45ec45-1ef2-46e6-bf33-0c092885679f"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.432916 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfksc\" (UniqueName: \"kubernetes.io/projected/85fbfb01-ec16-4977-9f54-16b0916b93ca-kube-api-access-xfksc\") pod \"ceilometer-0\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.506141 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.515910 4750 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef45ec45-1ef2-46e6-bf33-0c092885679f-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.641315 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j4pfx" event={"ID":"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6","Type":"ContainerStarted","Data":"78cc53e3a4300b13510980d34453082e4103251802af1436543224ed626b17b8"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.641377 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j4pfx" event={"ID":"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6","Type":"ContainerStarted","Data":"9e812e11ce85db367c1cd4363c9647c2ab1fb9808c82be36fe0c3ffe0042c88c"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.656192 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" event={"ID":"55185786-16ac-45a0-90b2-e44d0e833c23","Type":"ContainerStarted","Data":"556664ac95ccd0195f8e2bbd01932d6e568526ade20a2dc2eba689eda8675b68"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.668582 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-j4pfx" podStartSLOduration=7.668565106 podStartE2EDuration="7.668565106s" podCreationTimestamp="2026-03-09 18:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:41.663757526 +0000 UTC m=+1403.006229924" watchObservedRunningTime="2026-03-09 18:48:41.668565106 +0000 UTC m=+1403.011037504" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.673564 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qn2pj" event={"ID":"e3837950-d435-4f57-9bde-3c5172cba1d8","Type":"ContainerStarted","Data":"3628c72665cdec94060f5e21f9965583db46f67b888c64cc7af16cd84ad4c2a6"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.673608 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qn2pj" event={"ID":"e3837950-d435-4f57-9bde-3c5172cba1d8","Type":"ContainerStarted","Data":"bbfe1aadc3fba1eecf4063c31105a366492b93cd71efb205553c9c513c46a741"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.692433 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ddc4b4b8d-f5q7t" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.692756 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ddc4b4b8d-f5q7t" event={"ID":"ef45ec45-1ef2-46e6-bf33-0c092885679f","Type":"ContainerDied","Data":"47c7d7888f925219c0f83327043d08e690f53d43cb175c3af95161e414de0531"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.692823 4750 scope.go:117] "RemoveContainer" containerID="8f715bad4ddba91fed463c55ef1bb23df16777368f645f8c1a8928028b49eb87" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.699479 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" podStartSLOduration=7.699465935 podStartE2EDuration="7.699465935s" podCreationTimestamp="2026-03-09 18:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:41.686951285 +0000 UTC m=+1403.029423673" watchObservedRunningTime="2026-03-09 18:48:41.699465935 +0000 UTC m=+1403.041938333" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.705240 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" event={"ID":"f9746db5-1458-4326-ba1f-5b3abbecce78","Type":"ContainerStarted","Data":"0a9740590b05e48961606c6388f58f7ec58ddafc85b5662202bb308aca94c9d0"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.705283 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" event={"ID":"f9746db5-1458-4326-ba1f-5b3abbecce78","Type":"ContainerStarted","Data":"c61ed5ff9dd7b38d8f5ca793e30d40ad725e369a899429ae883177f25a9b634b"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.715427 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dk8lh" event={"ID":"e3b2776f-8175-48d4-8c86-9269df88aa64","Type":"ContainerStarted","Data":"9a8666f2d050f4143bc6b8c4f1b6a39f930ca1b8ba111b6cd701d0e899769dd6"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.715486 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dk8lh" event={"ID":"e3b2776f-8175-48d4-8c86-9269df88aa64","Type":"ContainerStarted","Data":"95faf692d3ac575e4a368ced25350963e3613e904fe9e067a2d8adad92872061"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.719688 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-08df-account-create-update-9t8fq" event={"ID":"081630cf-500b-49f1-992d-233ed78ff1bc","Type":"ContainerStarted","Data":"d6cb1251bdad02583497704b7ed6a8c6e3a45e2134af935927f1aad66508597f"} Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.726346 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-qn2pj" podStartSLOduration=7.726324705 podStartE2EDuration="7.726324705s" podCreationTimestamp="2026-03-09 18:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:41.704879053 +0000 UTC m=+1403.047351451" watchObservedRunningTime="2026-03-09 18:48:41.726324705 +0000 UTC m=+1403.068797103" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.768105 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" podStartSLOduration=7.76808602 podStartE2EDuration="7.76808602s" podCreationTimestamp="2026-03-09 18:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:41.736131371 +0000 UTC m=+1403.078603769" watchObservedRunningTime="2026-03-09 18:48:41.76808602 +0000 UTC m=+1403.110558408" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.795001 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-08df-account-create-update-9t8fq" podStartSLOduration=7.7949820899999995 podStartE2EDuration="7.79498209s" podCreationTimestamp="2026-03-09 18:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:41.776552729 +0000 UTC m=+1403.119025127" watchObservedRunningTime="2026-03-09 18:48:41.79498209 +0000 UTC m=+1403.137454488" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.872681 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:41 crc kubenswrapper[4750]: I0309 18:48:41.884597 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-54cf6684f5-bz5sx" Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.004048 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5ddc4b4b8d-f5q7t"] Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.017817 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5ddc4b4b8d-f5q7t"] Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.077944 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.162689 4750 scope.go:117] "RemoveContainer" containerID="7c882d7feb763fa768bb968ae79a038f49e22c03454de50f1ddfe39b393e082b" Mar 09 18:48:42 crc kubenswrapper[4750]: W0309 18:48:42.181449 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85fbfb01_ec16_4977_9f54_16b0916b93ca.slice/crio-5b22d94a11fa2044b6629e23108a7b22ca7145697895d841cc540028310a8946 WatchSource:0}: Error finding container 5b22d94a11fa2044b6629e23108a7b22ca7145697895d841cc540028310a8946: Status 404 returned error can't find the container with id 5b22d94a11fa2044b6629e23108a7b22ca7145697895d841cc540028310a8946 Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.730585 4750 generic.go:334] "Generic (PLEG): container finished" podID="e3837950-d435-4f57-9bde-3c5172cba1d8" containerID="3628c72665cdec94060f5e21f9965583db46f67b888c64cc7af16cd84ad4c2a6" exitCode=0 Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.730684 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qn2pj" event={"ID":"e3837950-d435-4f57-9bde-3c5172cba1d8","Type":"ContainerDied","Data":"3628c72665cdec94060f5e21f9965583db46f67b888c64cc7af16cd84ad4c2a6"} Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.733918 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerStarted","Data":"a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7"} Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.733968 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerStarted","Data":"5b22d94a11fa2044b6629e23108a7b22ca7145697895d841cc540028310a8946"} Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.738228 4750 generic.go:334] "Generic (PLEG): container finished" podID="081630cf-500b-49f1-992d-233ed78ff1bc" containerID="d6cb1251bdad02583497704b7ed6a8c6e3a45e2134af935927f1aad66508597f" exitCode=0 Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.738301 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-08df-account-create-update-9t8fq" event={"ID":"081630cf-500b-49f1-992d-233ed78ff1bc","Type":"ContainerDied","Data":"d6cb1251bdad02583497704b7ed6a8c6e3a45e2134af935927f1aad66508597f"} Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.741318 4750 generic.go:334] "Generic (PLEG): container finished" podID="f9746db5-1458-4326-ba1f-5b3abbecce78" containerID="0a9740590b05e48961606c6388f58f7ec58ddafc85b5662202bb308aca94c9d0" exitCode=0 Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.741380 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" event={"ID":"f9746db5-1458-4326-ba1f-5b3abbecce78","Type":"ContainerDied","Data":"0a9740590b05e48961606c6388f58f7ec58ddafc85b5662202bb308aca94c9d0"} Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.743772 4750 generic.go:334] "Generic (PLEG): container finished" podID="32248ec3-5f60-4ded-b77d-b9e88b7dcfe6" containerID="78cc53e3a4300b13510980d34453082e4103251802af1436543224ed626b17b8" exitCode=0 Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.743888 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j4pfx" event={"ID":"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6","Type":"ContainerDied","Data":"78cc53e3a4300b13510980d34453082e4103251802af1436543224ed626b17b8"} Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.747304 4750 generic.go:334] "Generic (PLEG): container finished" podID="55185786-16ac-45a0-90b2-e44d0e833c23" containerID="556664ac95ccd0195f8e2bbd01932d6e568526ade20a2dc2eba689eda8675b68" exitCode=0 Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.747363 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" event={"ID":"55185786-16ac-45a0-90b2-e44d0e833c23","Type":"ContainerDied","Data":"556664ac95ccd0195f8e2bbd01932d6e568526ade20a2dc2eba689eda8675b68"} Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.757252 4750 generic.go:334] "Generic (PLEG): container finished" podID="e3b2776f-8175-48d4-8c86-9269df88aa64" containerID="9a8666f2d050f4143bc6b8c4f1b6a39f930ca1b8ba111b6cd701d0e899769dd6" exitCode=0 Mar 09 18:48:42 crc kubenswrapper[4750]: I0309 18:48:42.757346 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dk8lh" event={"ID":"e3b2776f-8175-48d4-8c86-9269df88aa64","Type":"ContainerDied","Data":"9a8666f2d050f4143bc6b8c4f1b6a39f930ca1b8ba111b6cd701d0e899769dd6"} Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.243038 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.366033 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-776r8\" (UniqueName: \"kubernetes.io/projected/e3b2776f-8175-48d4-8c86-9269df88aa64-kube-api-access-776r8\") pod \"e3b2776f-8175-48d4-8c86-9269df88aa64\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.366297 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b2776f-8175-48d4-8c86-9269df88aa64-operator-scripts\") pod \"e3b2776f-8175-48d4-8c86-9269df88aa64\" (UID: \"e3b2776f-8175-48d4-8c86-9269df88aa64\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.367685 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3b2776f-8175-48d4-8c86-9269df88aa64-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e3b2776f-8175-48d4-8c86-9269df88aa64" (UID: "e3b2776f-8175-48d4-8c86-9269df88aa64"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.369954 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3b2776f-8175-48d4-8c86-9269df88aa64-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.374928 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b2776f-8175-48d4-8c86-9269df88aa64-kube-api-access-776r8" (OuterVolumeSpecName: "kube-api-access-776r8") pod "e3b2776f-8175-48d4-8c86-9269df88aa64" (UID: "e3b2776f-8175-48d4-8c86-9269df88aa64"). InnerVolumeSpecName "kube-api-access-776r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.439996 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef45ec45-1ef2-46e6-bf33-0c092885679f" path="/var/lib/kubelet/pods/ef45ec45-1ef2-46e6-bf33-0c092885679f/volumes" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.471948 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-776r8\" (UniqueName: \"kubernetes.io/projected/e3b2776f-8175-48d4-8c86-9269df88aa64-kube-api-access-776r8\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.474887 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.573711 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tpqw\" (UniqueName: \"kubernetes.io/projected/5fe67bbc-847f-43b2-878c-13f9b54f10c6-kube-api-access-6tpqw\") pod \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.573891 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-combined-ca-bundle\") pod \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.573927 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data\") pod \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.574009 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data-custom\") pod \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.574037 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fe67bbc-847f-43b2-878c-13f9b54f10c6-etc-machine-id\") pod \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.574051 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-scripts\") pod \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.574089 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fe67bbc-847f-43b2-878c-13f9b54f10c6-logs\") pod \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\" (UID: \"5fe67bbc-847f-43b2-878c-13f9b54f10c6\") " Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.574908 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5fe67bbc-847f-43b2-878c-13f9b54f10c6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5fe67bbc-847f-43b2-878c-13f9b54f10c6" (UID: "5fe67bbc-847f-43b2-878c-13f9b54f10c6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.575011 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe67bbc-847f-43b2-878c-13f9b54f10c6-logs" (OuterVolumeSpecName: "logs") pod "5fe67bbc-847f-43b2-878c-13f9b54f10c6" (UID: "5fe67bbc-847f-43b2-878c-13f9b54f10c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.579876 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-scripts" (OuterVolumeSpecName: "scripts") pod "5fe67bbc-847f-43b2-878c-13f9b54f10c6" (UID: "5fe67bbc-847f-43b2-878c-13f9b54f10c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.584691 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5fe67bbc-847f-43b2-878c-13f9b54f10c6" (UID: "5fe67bbc-847f-43b2-878c-13f9b54f10c6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.586607 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe67bbc-847f-43b2-878c-13f9b54f10c6-kube-api-access-6tpqw" (OuterVolumeSpecName: "kube-api-access-6tpqw") pod "5fe67bbc-847f-43b2-878c-13f9b54f10c6" (UID: "5fe67bbc-847f-43b2-878c-13f9b54f10c6"). InnerVolumeSpecName "kube-api-access-6tpqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.639569 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fe67bbc-847f-43b2-878c-13f9b54f10c6" (UID: "5fe67bbc-847f-43b2-878c-13f9b54f10c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.646874 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data" (OuterVolumeSpecName: "config-data") pod "5fe67bbc-847f-43b2-878c-13f9b54f10c6" (UID: "5fe67bbc-847f-43b2-878c-13f9b54f10c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.677339 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tpqw\" (UniqueName: \"kubernetes.io/projected/5fe67bbc-847f-43b2-878c-13f9b54f10c6-kube-api-access-6tpqw\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.677393 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.677408 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.677421 4750 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.677433 4750 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fe67bbc-847f-43b2-878c-13f9b54f10c6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.677444 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fe67bbc-847f-43b2-878c-13f9b54f10c6-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.677457 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fe67bbc-847f-43b2-878c-13f9b54f10c6-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.781390 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dk8lh" event={"ID":"e3b2776f-8175-48d4-8c86-9269df88aa64","Type":"ContainerDied","Data":"95faf692d3ac575e4a368ced25350963e3613e904fe9e067a2d8adad92872061"} Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.781787 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95faf692d3ac575e4a368ced25350963e3613e904fe9e067a2d8adad92872061" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.781850 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dk8lh" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.785278 4750 generic.go:334] "Generic (PLEG): container finished" podID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerID="cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492" exitCode=137 Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.785362 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fe67bbc-847f-43b2-878c-13f9b54f10c6","Type":"ContainerDied","Data":"cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492"} Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.785395 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.785504 4750 scope.go:117] "RemoveContainer" containerID="cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.785480 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fe67bbc-847f-43b2-878c-13f9b54f10c6","Type":"ContainerDied","Data":"4af08225eb472430343edcac649fb94efba0180d48ce0a8ea35af161880244c5"} Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.788452 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerStarted","Data":"22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c"} Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.788488 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerStarted","Data":"a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e"} Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.841024 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.879448 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.901656 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.911836 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:43 crc kubenswrapper[4750]: E0309 18:48:43.912342 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b2776f-8175-48d4-8c86-9269df88aa64" containerName="mariadb-database-create" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.912361 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b2776f-8175-48d4-8c86-9269df88aa64" containerName="mariadb-database-create" Mar 09 18:48:43 crc kubenswrapper[4750]: E0309 18:48:43.912384 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api-log" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.912392 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api-log" Mar 09 18:48:43 crc kubenswrapper[4750]: E0309 18:48:43.912410 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.912416 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.912610 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api-log" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.912626 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" containerName="cinder-api" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.912640 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b2776f-8175-48d4-8c86-9269df88aa64" containerName="mariadb-database-create" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.913899 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.918227 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.918242 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.918402 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.922282 4750 scope.go:117] "RemoveContainer" containerID="3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf" Mar 09 18:48:43 crc kubenswrapper[4750]: I0309 18:48:43.923236 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.008219 4750 scope.go:117] "RemoveContainer" containerID="cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492" Mar 09 18:48:44 crc kubenswrapper[4750]: E0309 18:48:44.008611 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492\": container with ID starting with cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492 not found: ID does not exist" containerID="cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.008735 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492"} err="failed to get container status \"cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492\": rpc error: code = NotFound desc = could not find container \"cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492\": container with ID starting with cdbfebe012f0c85ba6f0f5b819796bbc83b2f9a52925e14f2ef6f94516776492 not found: ID does not exist" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.008821 4750 scope.go:117] "RemoveContainer" containerID="3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf" Mar 09 18:48:44 crc kubenswrapper[4750]: E0309 18:48:44.011359 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf\": container with ID starting with 3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf not found: ID does not exist" containerID="3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.011398 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf"} err="failed to get container status \"3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf\": rpc error: code = NotFound desc = could not find container \"3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf\": container with ID starting with 3ec28cc6639d932df0ec0ff17d7fea9d95fc08393de287e158318fa9edae25bf not found: ID does not exist" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093368 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093528 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thn6s\" (UniqueName: \"kubernetes.io/projected/af01db09-e910-4f84-87d7-d8565dd13633-kube-api-access-thn6s\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093552 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-public-tls-certs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093579 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093604 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af01db09-e910-4f84-87d7-d8565dd13633-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093625 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-config-data\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093678 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-scripts\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093712 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-config-data-custom\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.093792 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af01db09-e910-4f84-87d7-d8565dd13633-logs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.202871 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thn6s\" (UniqueName: \"kubernetes.io/projected/af01db09-e910-4f84-87d7-d8565dd13633-kube-api-access-thn6s\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.202916 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-public-tls-certs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.202942 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.202964 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af01db09-e910-4f84-87d7-d8565dd13633-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.202987 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-config-data\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.203019 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-scripts\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.203053 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-config-data-custom\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.203087 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af01db09-e910-4f84-87d7-d8565dd13633-logs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.203165 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.203471 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af01db09-e910-4f84-87d7-d8565dd13633-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.205108 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af01db09-e910-4f84-87d7-d8565dd13633-logs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.220216 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.220256 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.220708 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-config-data-custom\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.222825 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-scripts\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.224313 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-config-data\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.227264 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thn6s\" (UniqueName: \"kubernetes.io/projected/af01db09-e910-4f84-87d7-d8565dd13633-kube-api-access-thn6s\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.234830 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af01db09-e910-4f84-87d7-d8565dd13633-public-tls-certs\") pod \"cinder-api-0\" (UID: \"af01db09-e910-4f84-87d7-d8565dd13633\") " pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.275326 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.439208 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.612793 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081630cf-500b-49f1-992d-233ed78ff1bc-operator-scripts\") pod \"081630cf-500b-49f1-992d-233ed78ff1bc\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.612971 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9rb9\" (UniqueName: \"kubernetes.io/projected/081630cf-500b-49f1-992d-233ed78ff1bc-kube-api-access-c9rb9\") pod \"081630cf-500b-49f1-992d-233ed78ff1bc\" (UID: \"081630cf-500b-49f1-992d-233ed78ff1bc\") " Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.614786 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/081630cf-500b-49f1-992d-233ed78ff1bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "081630cf-500b-49f1-992d-233ed78ff1bc" (UID: "081630cf-500b-49f1-992d-233ed78ff1bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.622958 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081630cf-500b-49f1-992d-233ed78ff1bc-kube-api-access-c9rb9" (OuterVolumeSpecName: "kube-api-access-c9rb9") pod "081630cf-500b-49f1-992d-233ed78ff1bc" (UID: "081630cf-500b-49f1-992d-233ed78ff1bc"). InnerVolumeSpecName "kube-api-access-c9rb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.718130 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081630cf-500b-49f1-992d-233ed78ff1bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.718161 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9rb9\" (UniqueName: \"kubernetes.io/projected/081630cf-500b-49f1-992d-233ed78ff1bc-kube-api-access-c9rb9\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.813939 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" event={"ID":"55185786-16ac-45a0-90b2-e44d0e833c23","Type":"ContainerDied","Data":"bad2db5dc077a6a7a701138b560c6a587839aab4ebbe118efe29004531375f67"} Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.813983 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bad2db5dc077a6a7a701138b560c6a587839aab4ebbe118efe29004531375f67" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.828543 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-08df-account-create-update-9t8fq" event={"ID":"081630cf-500b-49f1-992d-233ed78ff1bc","Type":"ContainerDied","Data":"3d3c7199459b04c1ac5ae9e1a2917ca8ace837a6c02349d3f116adb4177bf53f"} Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.828587 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d3c7199459b04c1ac5ae9e1a2917ca8ace837a6c02349d3f116adb4177bf53f" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.828684 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-08df-account-create-update-9t8fq" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.878196 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" event={"ID":"f9746db5-1458-4326-ba1f-5b3abbecce78","Type":"ContainerDied","Data":"c61ed5ff9dd7b38d8f5ca793e30d40ad725e369a899429ae883177f25a9b634b"} Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.878272 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c61ed5ff9dd7b38d8f5ca793e30d40ad725e369a899429ae883177f25a9b634b" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.900001 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j4pfx" event={"ID":"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6","Type":"ContainerDied","Data":"9e812e11ce85db367c1cd4363c9647c2ab1fb9808c82be36fe0c3ffe0042c88c"} Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.900041 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e812e11ce85db367c1cd4363c9647c2ab1fb9808c82be36fe0c3ffe0042c88c" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.922249 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.950877 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:44 crc kubenswrapper[4750]: I0309 18:48:44.999020 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.021290 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.035317 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwjvx\" (UniqueName: \"kubernetes.io/projected/f9746db5-1458-4326-ba1f-5b3abbecce78-kube-api-access-mwjvx\") pod \"f9746db5-1458-4326-ba1f-5b3abbecce78\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.035576 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9746db5-1458-4326-ba1f-5b3abbecce78-operator-scripts\") pod \"f9746db5-1458-4326-ba1f-5b3abbecce78\" (UID: \"f9746db5-1458-4326-ba1f-5b3abbecce78\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.036566 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9746db5-1458-4326-ba1f-5b3abbecce78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f9746db5-1458-4326-ba1f-5b3abbecce78" (UID: "f9746db5-1458-4326-ba1f-5b3abbecce78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.069101 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9746db5-1458-4326-ba1f-5b3abbecce78-kube-api-access-mwjvx" (OuterVolumeSpecName: "kube-api-access-mwjvx") pod "f9746db5-1458-4326-ba1f-5b3abbecce78" (UID: "f9746db5-1458-4326-ba1f-5b3abbecce78"). InnerVolumeSpecName "kube-api-access-mwjvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.139163 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqm6j\" (UniqueName: \"kubernetes.io/projected/e3837950-d435-4f57-9bde-3c5172cba1d8-kube-api-access-jqm6j\") pod \"e3837950-d435-4f57-9bde-3c5172cba1d8\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.139199 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bghp4\" (UniqueName: \"kubernetes.io/projected/55185786-16ac-45a0-90b2-e44d0e833c23-kube-api-access-bghp4\") pod \"55185786-16ac-45a0-90b2-e44d0e833c23\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.139272 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-operator-scripts\") pod \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.139322 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3837950-d435-4f57-9bde-3c5172cba1d8-operator-scripts\") pod \"e3837950-d435-4f57-9bde-3c5172cba1d8\" (UID: \"e3837950-d435-4f57-9bde-3c5172cba1d8\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.139384 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2vq2\" (UniqueName: \"kubernetes.io/projected/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-kube-api-access-h2vq2\") pod \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\" (UID: \"32248ec3-5f60-4ded-b77d-b9e88b7dcfe6\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.139493 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55185786-16ac-45a0-90b2-e44d0e833c23-operator-scripts\") pod \"55185786-16ac-45a0-90b2-e44d0e833c23\" (UID: \"55185786-16ac-45a0-90b2-e44d0e833c23\") " Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.140352 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwjvx\" (UniqueName: \"kubernetes.io/projected/f9746db5-1458-4326-ba1f-5b3abbecce78-kube-api-access-mwjvx\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.140391 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9746db5-1458-4326-ba1f-5b3abbecce78-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.140881 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55185786-16ac-45a0-90b2-e44d0e833c23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55185786-16ac-45a0-90b2-e44d0e833c23" (UID: "55185786-16ac-45a0-90b2-e44d0e833c23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.141891 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3837950-d435-4f57-9bde-3c5172cba1d8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e3837950-d435-4f57-9bde-3c5172cba1d8" (UID: "e3837950-d435-4f57-9bde-3c5172cba1d8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.144976 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32248ec3-5f60-4ded-b77d-b9e88b7dcfe6" (UID: "32248ec3-5f60-4ded-b77d-b9e88b7dcfe6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.152027 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3837950-d435-4f57-9bde-3c5172cba1d8-kube-api-access-jqm6j" (OuterVolumeSpecName: "kube-api-access-jqm6j") pod "e3837950-d435-4f57-9bde-3c5172cba1d8" (UID: "e3837950-d435-4f57-9bde-3c5172cba1d8"). InnerVolumeSpecName "kube-api-access-jqm6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.156686 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-kube-api-access-h2vq2" (OuterVolumeSpecName: "kube-api-access-h2vq2") pod "32248ec3-5f60-4ded-b77d-b9e88b7dcfe6" (UID: "32248ec3-5f60-4ded-b77d-b9e88b7dcfe6"). InnerVolumeSpecName "kube-api-access-h2vq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.158067 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55185786-16ac-45a0-90b2-e44d0e833c23-kube-api-access-bghp4" (OuterVolumeSpecName: "kube-api-access-bghp4") pod "55185786-16ac-45a0-90b2-e44d0e833c23" (UID: "55185786-16ac-45a0-90b2-e44d0e833c23"). InnerVolumeSpecName "kube-api-access-bghp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.221023 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.251808 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.252122 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3837950-d435-4f57-9bde-3c5172cba1d8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.252178 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2vq2\" (UniqueName: \"kubernetes.io/projected/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6-kube-api-access-h2vq2\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.252242 4750 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55185786-16ac-45a0-90b2-e44d0e833c23-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.252334 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqm6j\" (UniqueName: \"kubernetes.io/projected/e3837950-d435-4f57-9bde-3c5172cba1d8-kube-api-access-jqm6j\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.252411 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bghp4\" (UniqueName: \"kubernetes.io/projected/55185786-16ac-45a0-90b2-e44d0e833c23-kube-api-access-bghp4\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.389005 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe67bbc-847f-43b2-878c-13f9b54f10c6" path="/var/lib/kubelet/pods/5fe67bbc-847f-43b2-878c-13f9b54f10c6/volumes" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.919539 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af01db09-e910-4f84-87d7-d8565dd13633","Type":"ContainerStarted","Data":"ba4d780baa6318214be7a3740ca24664e5d5b15f8024bfcebcb1ed0a1282df2f"} Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.926184 4750 generic.go:334] "Generic (PLEG): container finished" podID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerID="5ad8d0972171dc5644b0c875da08f240816d847855e0e2602a5e8575930a3e3e" exitCode=0 Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.926274 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6db5794d-f9pxj" event={"ID":"9d5b4b2b-6058-44ac-b24b-e5ef2423f830","Type":"ContainerDied","Data":"5ad8d0972171dc5644b0c875da08f240816d847855e0e2602a5e8575930a3e3e"} Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.926334 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6db5794d-f9pxj" event={"ID":"9d5b4b2b-6058-44ac-b24b-e5ef2423f830","Type":"ContainerDied","Data":"779e4b67cda39bfa39b4708149b50445e378fd8f66e9116a49fc6b8097ea96b5"} Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.926347 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="779e4b67cda39bfa39b4708149b50445e378fd8f66e9116a49fc6b8097ea96b5" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.927772 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b786-account-create-update-dr9fd" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.927819 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qn2pj" event={"ID":"e3837950-d435-4f57-9bde-3c5172cba1d8","Type":"ContainerDied","Data":"bbfe1aadc3fba1eecf4063c31105a366492b93cd71efb205553c9c513c46a741"} Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.927846 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbfe1aadc3fba1eecf4063c31105a366492b93cd71efb205553c9c513c46a741" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.928092 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f67-account-create-update-l2t7v" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.928148 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qn2pj" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.928298 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j4pfx" Mar 09 18:48:45 crc kubenswrapper[4750]: I0309 18:48:45.996943 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.070272 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8rz4\" (UniqueName: \"kubernetes.io/projected/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-kube-api-access-z8rz4\") pod \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.070329 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-combined-ca-bundle\") pod \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.070354 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-config\") pod \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.070380 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-ovndb-tls-certs\") pod \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.070462 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-httpd-config\") pod \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\" (UID: \"9d5b4b2b-6058-44ac-b24b-e5ef2423f830\") " Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.102335 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-kube-api-access-z8rz4" (OuterVolumeSpecName: "kube-api-access-z8rz4") pod "9d5b4b2b-6058-44ac-b24b-e5ef2423f830" (UID: "9d5b4b2b-6058-44ac-b24b-e5ef2423f830"). InnerVolumeSpecName "kube-api-access-z8rz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.114002 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9d5b4b2b-6058-44ac-b24b-e5ef2423f830" (UID: "9d5b4b2b-6058-44ac-b24b-e5ef2423f830"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.172866 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8rz4\" (UniqueName: \"kubernetes.io/projected/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-kube-api-access-z8rz4\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.172911 4750 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.288916 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d5b4b2b-6058-44ac-b24b-e5ef2423f830" (UID: "9d5b4b2b-6058-44ac-b24b-e5ef2423f830"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.380887 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.381841 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-config" (OuterVolumeSpecName: "config") pod "9d5b4b2b-6058-44ac-b24b-e5ef2423f830" (UID: "9d5b4b2b-6058-44ac-b24b-e5ef2423f830"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.388793 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9d5b4b2b-6058-44ac-b24b-e5ef2423f830" (UID: "9d5b4b2b-6058-44ac-b24b-e5ef2423f830"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.484473 4750 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.484511 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d5b4b2b-6058-44ac-b24b-e5ef2423f830-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.952728 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerStarted","Data":"7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c"} Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.952910 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-central-agent" containerID="cri-o://a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7" gracePeriod=30 Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.953028 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.953870 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="proxy-httpd" containerID="cri-o://7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c" gracePeriod=30 Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.953919 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="sg-core" containerID="cri-o://22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c" gracePeriod=30 Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.953958 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-notification-agent" containerID="cri-o://a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e" gracePeriod=30 Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.972053 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d6db5794d-f9pxj" Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.973954 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af01db09-e910-4f84-87d7-d8565dd13633","Type":"ContainerStarted","Data":"721618a76947bd40414049f0620119065b8d44fc8947bf5aa56f89d43e8d7278"} Mar 09 18:48:46 crc kubenswrapper[4750]: I0309 18:48:46.991519 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.340498039 podStartE2EDuration="5.99150249s" podCreationTimestamp="2026-03-09 18:48:41 +0000 UTC" firstStartedPulling="2026-03-09 18:48:42.187214626 +0000 UTC m=+1403.529687024" lastFinishedPulling="2026-03-09 18:48:45.838219067 +0000 UTC m=+1407.180691475" observedRunningTime="2026-03-09 18:48:46.980622904 +0000 UTC m=+1408.323095302" watchObservedRunningTime="2026-03-09 18:48:46.99150249 +0000 UTC m=+1408.333974888" Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.017875 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d6db5794d-f9pxj"] Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.028557 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d6db5794d-f9pxj"] Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.385062 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" path="/var/lib/kubelet/pods/9d5b4b2b-6058-44ac-b24b-e5ef2423f830/volumes" Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.982842 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af01db09-e910-4f84-87d7-d8565dd13633","Type":"ContainerStarted","Data":"55b96f4a50eb61e0a897865d2ac7e3f562798cd41bff042172a6322326601ea7"} Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.983002 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.986147 4750 generic.go:334] "Generic (PLEG): container finished" podID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerID="7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c" exitCode=0 Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.986197 4750 generic.go:334] "Generic (PLEG): container finished" podID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerID="22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c" exitCode=2 Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.986205 4750 generic.go:334] "Generic (PLEG): container finished" podID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerID="a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e" exitCode=0 Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.986225 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerDied","Data":"7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c"} Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.986273 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerDied","Data":"22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c"} Mar 09 18:48:47 crc kubenswrapper[4750]: I0309 18:48:47.986284 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerDied","Data":"a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e"} Mar 09 18:48:48 crc kubenswrapper[4750]: I0309 18:48:48.010251 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.010231467 podStartE2EDuration="5.010231467s" podCreationTimestamp="2026-03-09 18:48:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:48:48.001466649 +0000 UTC m=+1409.343939057" watchObservedRunningTime="2026-03-09 18:48:48.010231467 +0000 UTC m=+1409.352703865" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.731692 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.777807 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfksc\" (UniqueName: \"kubernetes.io/projected/85fbfb01-ec16-4977-9f54-16b0916b93ca-kube-api-access-xfksc\") pod \"85fbfb01-ec16-4977-9f54-16b0916b93ca\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.777936 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-log-httpd\") pod \"85fbfb01-ec16-4977-9f54-16b0916b93ca\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.777979 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-config-data\") pod \"85fbfb01-ec16-4977-9f54-16b0916b93ca\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.778045 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-combined-ca-bundle\") pod \"85fbfb01-ec16-4977-9f54-16b0916b93ca\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.778084 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-sg-core-conf-yaml\") pod \"85fbfb01-ec16-4977-9f54-16b0916b93ca\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.778204 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-scripts\") pod \"85fbfb01-ec16-4977-9f54-16b0916b93ca\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.778236 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-run-httpd\") pod \"85fbfb01-ec16-4977-9f54-16b0916b93ca\" (UID: \"85fbfb01-ec16-4977-9f54-16b0916b93ca\") " Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.779040 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "85fbfb01-ec16-4977-9f54-16b0916b93ca" (UID: "85fbfb01-ec16-4977-9f54-16b0916b93ca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.780337 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "85fbfb01-ec16-4977-9f54-16b0916b93ca" (UID: "85fbfb01-ec16-4977-9f54-16b0916b93ca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.788248 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85fbfb01-ec16-4977-9f54-16b0916b93ca-kube-api-access-xfksc" (OuterVolumeSpecName: "kube-api-access-xfksc") pod "85fbfb01-ec16-4977-9f54-16b0916b93ca" (UID: "85fbfb01-ec16-4977-9f54-16b0916b93ca"). InnerVolumeSpecName "kube-api-access-xfksc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.792169 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-scripts" (OuterVolumeSpecName: "scripts") pod "85fbfb01-ec16-4977-9f54-16b0916b93ca" (UID: "85fbfb01-ec16-4977-9f54-16b0916b93ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.877446 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "85fbfb01-ec16-4977-9f54-16b0916b93ca" (UID: "85fbfb01-ec16-4977-9f54-16b0916b93ca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.880531 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfksc\" (UniqueName: \"kubernetes.io/projected/85fbfb01-ec16-4977-9f54-16b0916b93ca-kube-api-access-xfksc\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.880568 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.880582 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.880594 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.880606 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85fbfb01-ec16-4977-9f54-16b0916b93ca-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.946866 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85fbfb01-ec16-4977-9f54-16b0916b93ca" (UID: "85fbfb01-ec16-4977-9f54-16b0916b93ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.958922 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-config-data" (OuterVolumeSpecName: "config-data") pod "85fbfb01-ec16-4977-9f54-16b0916b93ca" (UID: "85fbfb01-ec16-4977-9f54-16b0916b93ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.982816 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.982847 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85fbfb01-ec16-4977-9f54-16b0916b93ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.982872 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5jn6"] Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983471 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32248ec3-5f60-4ded-b77d-b9e88b7dcfe6" containerName="mariadb-database-create" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983485 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="32248ec3-5f60-4ded-b77d-b9e88b7dcfe6" containerName="mariadb-database-create" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983502 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-httpd" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983509 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-httpd" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983519 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="proxy-httpd" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983526 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="proxy-httpd" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983546 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-notification-agent" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983552 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-notification-agent" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983559 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="sg-core" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983565 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="sg-core" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983573 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3837950-d435-4f57-9bde-3c5172cba1d8" containerName="mariadb-database-create" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983591 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3837950-d435-4f57-9bde-3c5172cba1d8" containerName="mariadb-database-create" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983601 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9746db5-1458-4326-ba1f-5b3abbecce78" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983607 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9746db5-1458-4326-ba1f-5b3abbecce78" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983619 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-api" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983625 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-api" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983863 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-central-agent" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983874 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-central-agent" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983889 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081630cf-500b-49f1-992d-233ed78ff1bc" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983895 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="081630cf-500b-49f1-992d-233ed78ff1bc" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: E0309 18:48:49.983902 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55185786-16ac-45a0-90b2-e44d0e833c23" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.983908 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="55185786-16ac-45a0-90b2-e44d0e833c23" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984116 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-api" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984127 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="proxy-httpd" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984134 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="sg-core" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984141 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="081630cf-500b-49f1-992d-233ed78ff1bc" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984151 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-central-agent" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984165 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3837950-d435-4f57-9bde-3c5172cba1d8" containerName="mariadb-database-create" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984173 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9746db5-1458-4326-ba1f-5b3abbecce78" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984184 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerName="ceilometer-notification-agent" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984191 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="55185786-16ac-45a0-90b2-e44d0e833c23" containerName="mariadb-account-create-update" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984198 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="32248ec3-5f60-4ded-b77d-b9e88b7dcfe6" containerName="mariadb-database-create" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.984212 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5b4b2b-6058-44ac-b24b-e5ef2423f830" containerName="neutron-httpd" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.985269 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.989350 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kljb7" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.989558 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 09 18:48:49 crc kubenswrapper[4750]: I0309 18:48:49.991541 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.013198 4750 generic.go:334] "Generic (PLEG): container finished" podID="85fbfb01-ec16-4977-9f54-16b0916b93ca" containerID="a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7" exitCode=0 Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.013765 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerDied","Data":"a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7"} Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.013807 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85fbfb01-ec16-4977-9f54-16b0916b93ca","Type":"ContainerDied","Data":"5b22d94a11fa2044b6629e23108a7b22ca7145697895d841cc540028310a8946"} Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.013831 4750 scope.go:117] "RemoveContainer" containerID="7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.014072 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.018257 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5jn6"] Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.043629 4750 scope.go:117] "RemoveContainer" containerID="22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.071015 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.075681 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.081766 4750 scope.go:117] "RemoveContainer" containerID="a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.084199 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-config-data\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.084454 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-scripts\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.084577 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.084706 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh9ds\" (UniqueName: \"kubernetes.io/projected/31e486b2-f819-4c3a-b7db-2262ad76bbb2-kube-api-access-wh9ds\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.097742 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.102206 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.112844 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.113237 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.130371 4750 scope.go:117] "RemoveContainer" containerID="a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.150784 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.164283 4750 scope.go:117] "RemoveContainer" containerID="7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c" Mar 09 18:48:50 crc kubenswrapper[4750]: E0309 18:48:50.167683 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c\": container with ID starting with 7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c not found: ID does not exist" containerID="7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.168035 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c"} err="failed to get container status \"7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c\": rpc error: code = NotFound desc = could not find container \"7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c\": container with ID starting with 7cd4cbb48af8bea623516b75345b3dd83206f6f4f6c68dc1608645b10a7eaf8c not found: ID does not exist" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.168181 4750 scope.go:117] "RemoveContainer" containerID="22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c" Mar 09 18:48:50 crc kubenswrapper[4750]: E0309 18:48:50.169241 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c\": container with ID starting with 22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c not found: ID does not exist" containerID="22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.169338 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c"} err="failed to get container status \"22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c\": rpc error: code = NotFound desc = could not find container \"22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c\": container with ID starting with 22c3092436c59cff0ed962847fc2a10462084171b2461582c610784ac717650c not found: ID does not exist" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.169434 4750 scope.go:117] "RemoveContainer" containerID="a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e" Mar 09 18:48:50 crc kubenswrapper[4750]: E0309 18:48:50.169732 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e\": container with ID starting with a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e not found: ID does not exist" containerID="a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.169844 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e"} err="failed to get container status \"a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e\": rpc error: code = NotFound desc = could not find container \"a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e\": container with ID starting with a95d89dbd296176411b508d570b52eea89ce6b169e36108369b0978562704e9e not found: ID does not exist" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.169962 4750 scope.go:117] "RemoveContainer" containerID="a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7" Mar 09 18:48:50 crc kubenswrapper[4750]: E0309 18:48:50.170222 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7\": container with ID starting with a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7 not found: ID does not exist" containerID="a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.170318 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7"} err="failed to get container status \"a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7\": rpc error: code = NotFound desc = could not find container \"a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7\": container with ID starting with a484975d332fa4df4bfb9076b01c5ca8dfad7b2afae4e3203293b869217d91e7 not found: ID does not exist" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.187177 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-scripts\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.187456 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.187555 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.187695 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.188168 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh9ds\" (UniqueName: \"kubernetes.io/projected/31e486b2-f819-4c3a-b7db-2262ad76bbb2-kube-api-access-wh9ds\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.188327 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctkqs\" (UniqueName: \"kubernetes.io/projected/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-kube-api-access-ctkqs\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.188464 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-run-httpd\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.188621 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-config-data\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.188804 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-scripts\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.188907 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-log-httpd\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.189016 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-config-data\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.194309 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-scripts\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.194324 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.198442 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-config-data\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.216797 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh9ds\" (UniqueName: \"kubernetes.io/projected/31e486b2-f819-4c3a-b7db-2262ad76bbb2-kube-api-access-wh9ds\") pod \"nova-cell0-conductor-db-sync-s5jn6\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.290913 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-config-data\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.290983 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-scripts\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.291007 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-log-httpd\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.291058 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.291082 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.291130 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctkqs\" (UniqueName: \"kubernetes.io/projected/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-kube-api-access-ctkqs\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.291165 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-run-httpd\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.291582 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-run-httpd\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.292027 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-log-httpd\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.295673 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-config-data\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.297913 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-scripts\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.306533 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.310451 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.321435 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.325365 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctkqs\" (UniqueName: \"kubernetes.io/projected/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-kube-api-access-ctkqs\") pod \"ceilometer-0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.430322 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:48:50 crc kubenswrapper[4750]: I0309 18:48:50.953768 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5jn6"] Mar 09 18:48:50 crc kubenswrapper[4750]: W0309 18:48:50.966895 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31e486b2_f819_4c3a_b7db_2262ad76bbb2.slice/crio-653f381241c4443d0373cddacc56767026eeffa91c73997c0081ba132d6875a2 WatchSource:0}: Error finding container 653f381241c4443d0373cddacc56767026eeffa91c73997c0081ba132d6875a2: Status 404 returned error can't find the container with id 653f381241c4443d0373cddacc56767026eeffa91c73997c0081ba132d6875a2 Mar 09 18:48:51 crc kubenswrapper[4750]: I0309 18:48:51.026916 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" event={"ID":"31e486b2-f819-4c3a-b7db-2262ad76bbb2","Type":"ContainerStarted","Data":"653f381241c4443d0373cddacc56767026eeffa91c73997c0081ba132d6875a2"} Mar 09 18:48:51 crc kubenswrapper[4750]: I0309 18:48:51.088656 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:51 crc kubenswrapper[4750]: I0309 18:48:51.373807 4750 scope.go:117] "RemoveContainer" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" Mar 09 18:48:51 crc kubenswrapper[4750]: E0309 18:48:51.374172 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(6796c4d9-5680-4c22-9ef0-815642b22c3f)\"" pod="openstack/watcher-decision-engine-0" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" Mar 09 18:48:51 crc kubenswrapper[4750]: I0309 18:48:51.392781 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85fbfb01-ec16-4977-9f54-16b0916b93ca" path="/var/lib/kubelet/pods/85fbfb01-ec16-4977-9f54-16b0916b93ca/volumes" Mar 09 18:48:52 crc kubenswrapper[4750]: I0309 18:48:52.038996 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerStarted","Data":"251326bb59b99bce352da45e99304905b7e42827e43b2e0cc6956960f2c05f77"} Mar 09 18:48:52 crc kubenswrapper[4750]: I0309 18:48:52.039298 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerStarted","Data":"185f2564b7cb9d28a2fb156eefc1a7e701b4f901fc25d867929dc91c6263f66e"} Mar 09 18:48:52 crc kubenswrapper[4750]: I0309 18:48:52.039311 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerStarted","Data":"a08b1a46bf2f708e37536358a40c0680c7bf53d07269d627ad140b450567d9c3"} Mar 09 18:48:53 crc kubenswrapper[4750]: I0309 18:48:53.049853 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerStarted","Data":"c95329c913aa4153a0df2980a7ad82897f19f7e2f5689fa287af480b69740bd7"} Mar 09 18:48:54 crc kubenswrapper[4750]: I0309 18:48:54.626315 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:48:55 crc kubenswrapper[4750]: I0309 18:48:55.086272 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerStarted","Data":"ae30cfac8180a25c9e2c053f5bacc9a01d5abe81180e7ac66ffb0cb06415a819"} Mar 09 18:48:55 crc kubenswrapper[4750]: I0309 18:48:55.086433 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:48:55 crc kubenswrapper[4750]: I0309 18:48:55.114340 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.003774564 podStartE2EDuration="5.114324813s" podCreationTimestamp="2026-03-09 18:48:50 +0000 UTC" firstStartedPulling="2026-03-09 18:48:51.103234638 +0000 UTC m=+1412.445707036" lastFinishedPulling="2026-03-09 18:48:54.213784887 +0000 UTC m=+1415.556257285" observedRunningTime="2026-03-09 18:48:55.113424548 +0000 UTC m=+1416.455896946" watchObservedRunningTime="2026-03-09 18:48:55.114324813 +0000 UTC m=+1416.456797211" Mar 09 18:48:56 crc kubenswrapper[4750]: I0309 18:48:56.097029 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-central-agent" containerID="cri-o://185f2564b7cb9d28a2fb156eefc1a7e701b4f901fc25d867929dc91c6263f66e" gracePeriod=30 Mar 09 18:48:56 crc kubenswrapper[4750]: I0309 18:48:56.097110 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="proxy-httpd" containerID="cri-o://ae30cfac8180a25c9e2c053f5bacc9a01d5abe81180e7ac66ffb0cb06415a819" gracePeriod=30 Mar 09 18:48:56 crc kubenswrapper[4750]: I0309 18:48:56.097219 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="sg-core" containerID="cri-o://c95329c913aa4153a0df2980a7ad82897f19f7e2f5689fa287af480b69740bd7" gracePeriod=30 Mar 09 18:48:56 crc kubenswrapper[4750]: I0309 18:48:56.097287 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-notification-agent" containerID="cri-o://251326bb59b99bce352da45e99304905b7e42827e43b2e0cc6956960f2c05f77" gracePeriod=30 Mar 09 18:48:56 crc kubenswrapper[4750]: I0309 18:48:56.652171 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107592 4750 generic.go:334] "Generic (PLEG): container finished" podID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerID="ae30cfac8180a25c9e2c053f5bacc9a01d5abe81180e7ac66ffb0cb06415a819" exitCode=0 Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107662 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerDied","Data":"ae30cfac8180a25c9e2c053f5bacc9a01d5abe81180e7ac66ffb0cb06415a819"} Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107689 4750 generic.go:334] "Generic (PLEG): container finished" podID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerID="c95329c913aa4153a0df2980a7ad82897f19f7e2f5689fa287af480b69740bd7" exitCode=2 Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107705 4750 generic.go:334] "Generic (PLEG): container finished" podID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerID="251326bb59b99bce352da45e99304905b7e42827e43b2e0cc6956960f2c05f77" exitCode=0 Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107887 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerDied","Data":"c95329c913aa4153a0df2980a7ad82897f19f7e2f5689fa287af480b69740bd7"} Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107901 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerDied","Data":"251326bb59b99bce352da45e99304905b7e42827e43b2e0cc6956960f2c05f77"} Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107912 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerDied","Data":"185f2564b7cb9d28a2fb156eefc1a7e701b4f901fc25d867929dc91c6263f66e"} Mar 09 18:48:57 crc kubenswrapper[4750]: I0309 18:48:57.107891 4750 generic.go:334] "Generic (PLEG): container finished" podID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerID="185f2564b7cb9d28a2fb156eefc1a7e701b4f901fc25d867929dc91c6263f66e" exitCode=0 Mar 09 18:48:58 crc kubenswrapper[4750]: I0309 18:48:58.991428 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:48:58 crc kubenswrapper[4750]: I0309 18:48:58.992590 4750 scope.go:117] "RemoveContainer" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" Mar 09 18:48:58 crc kubenswrapper[4750]: I0309 18:48:58.993067 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.566182 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588118 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-run-httpd\") pod \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588508 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctkqs\" (UniqueName: \"kubernetes.io/projected/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-kube-api-access-ctkqs\") pod \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588662 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-config-data\") pod \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588613 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4a8c3898-ecf7-41e7-81b5-af73d382d0d0" (UID: "4a8c3898-ecf7-41e7-81b5-af73d382d0d0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588791 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-combined-ca-bundle\") pod \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588920 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-scripts\") pod \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588968 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-sg-core-conf-yaml\") pod \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.588992 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-log-httpd\") pod \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\" (UID: \"4a8c3898-ecf7-41e7-81b5-af73d382d0d0\") " Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.589474 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4a8c3898-ecf7-41e7-81b5-af73d382d0d0" (UID: "4a8c3898-ecf7-41e7-81b5-af73d382d0d0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.590153 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.590168 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.612028 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-scripts" (OuterVolumeSpecName: "scripts") pod "4a8c3898-ecf7-41e7-81b5-af73d382d0d0" (UID: "4a8c3898-ecf7-41e7-81b5-af73d382d0d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.625135 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-kube-api-access-ctkqs" (OuterVolumeSpecName: "kube-api-access-ctkqs") pod "4a8c3898-ecf7-41e7-81b5-af73d382d0d0" (UID: "4a8c3898-ecf7-41e7-81b5-af73d382d0d0"). InnerVolumeSpecName "kube-api-access-ctkqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.638481 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4a8c3898-ecf7-41e7-81b5-af73d382d0d0" (UID: "4a8c3898-ecf7-41e7-81b5-af73d382d0d0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.692366 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.692566 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.692652 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctkqs\" (UniqueName: \"kubernetes.io/projected/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-kube-api-access-ctkqs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.737037 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a8c3898-ecf7-41e7-81b5-af73d382d0d0" (UID: "4a8c3898-ecf7-41e7-81b5-af73d382d0d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.800429 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-config-data" (OuterVolumeSpecName: "config-data") pod "4a8c3898-ecf7-41e7-81b5-af73d382d0d0" (UID: "4a8c3898-ecf7-41e7-81b5-af73d382d0d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.803533 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:02 crc kubenswrapper[4750]: I0309 18:49:02.803568 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8c3898-ecf7-41e7-81b5-af73d382d0d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.181240 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerStarted","Data":"e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95"} Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.184127 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" event={"ID":"31e486b2-f819-4c3a-b7db-2262ad76bbb2","Type":"ContainerStarted","Data":"7764ae635f169cc2206c2082440203a3b76a001e64c8eab35587e01281ad3419"} Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.188282 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a8c3898-ecf7-41e7-81b5-af73d382d0d0","Type":"ContainerDied","Data":"a08b1a46bf2f708e37536358a40c0680c7bf53d07269d627ad140b450567d9c3"} Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.188326 4750 scope.go:117] "RemoveContainer" containerID="ae30cfac8180a25c9e2c053f5bacc9a01d5abe81180e7ac66ffb0cb06415a819" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.188456 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.213858 4750 scope.go:117] "RemoveContainer" containerID="c95329c913aa4153a0df2980a7ad82897f19f7e2f5689fa287af480b69740bd7" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.232240 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" podStartSLOduration=2.785021242 podStartE2EDuration="14.232215942s" podCreationTimestamp="2026-03-09 18:48:49 +0000 UTC" firstStartedPulling="2026-03-09 18:48:50.971300314 +0000 UTC m=+1412.313772712" lastFinishedPulling="2026-03-09 18:49:02.418494994 +0000 UTC m=+1423.760967412" observedRunningTime="2026-03-09 18:49:03.22110816 +0000 UTC m=+1424.563580558" watchObservedRunningTime="2026-03-09 18:49:03.232215942 +0000 UTC m=+1424.574688340" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.242183 4750 scope.go:117] "RemoveContainer" containerID="251326bb59b99bce352da45e99304905b7e42827e43b2e0cc6956960f2c05f77" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.249277 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.261309 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.279529 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:03 crc kubenswrapper[4750]: E0309 18:49:03.280077 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="proxy-httpd" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280100 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="proxy-httpd" Mar 09 18:49:03 crc kubenswrapper[4750]: E0309 18:49:03.280128 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-notification-agent" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280137 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-notification-agent" Mar 09 18:49:03 crc kubenswrapper[4750]: E0309 18:49:03.280167 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="sg-core" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280173 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="sg-core" Mar 09 18:49:03 crc kubenswrapper[4750]: E0309 18:49:03.280184 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-central-agent" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280191 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-central-agent" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280428 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="proxy-httpd" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280458 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="sg-core" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280473 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-notification-agent" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.280486 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" containerName="ceilometer-central-agent" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.282401 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.287474 4750 scope.go:117] "RemoveContainer" containerID="185f2564b7cb9d28a2fb156eefc1a7e701b4f901fc25d867929dc91c6263f66e" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.287730 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.287866 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.312549 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.313599 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.313681 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-scripts\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.313729 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.313773 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-config-data\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.313816 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-log-httpd\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.313852 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-run-httpd\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.313924 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m85bz\" (UniqueName: \"kubernetes.io/projected/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-kube-api-access-m85bz\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.385796 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a8c3898-ecf7-41e7-81b5-af73d382d0d0" path="/var/lib/kubelet/pods/4a8c3898-ecf7-41e7-81b5-af73d382d0d0/volumes" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.416164 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m85bz\" (UniqueName: \"kubernetes.io/projected/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-kube-api-access-m85bz\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.416291 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.416376 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-scripts\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.416407 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.416461 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-config-data\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.416501 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-log-httpd\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.416539 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-run-httpd\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.417025 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-run-httpd\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.418468 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-log-httpd\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.421180 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.421793 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.422287 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-config-data\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.430319 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-scripts\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.434572 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m85bz\" (UniqueName: \"kubernetes.io/projected/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-kube-api-access-m85bz\") pod \"ceilometer-0\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " pod="openstack/ceilometer-0" Mar 09 18:49:03 crc kubenswrapper[4750]: I0309 18:49:03.610542 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:04 crc kubenswrapper[4750]: I0309 18:49:04.115525 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:04 crc kubenswrapper[4750]: W0309 18:49:04.119657 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podced5718e_bed8_44d0_a3e3_be97c80dc1d5.slice/crio-58a88d0572f6c186af4679d4d704bf3488b258adc08952d17236d4ce8fdcabf2 WatchSource:0}: Error finding container 58a88d0572f6c186af4679d4d704bf3488b258adc08952d17236d4ce8fdcabf2: Status 404 returned error can't find the container with id 58a88d0572f6c186af4679d4d704bf3488b258adc08952d17236d4ce8fdcabf2 Mar 09 18:49:04 crc kubenswrapper[4750]: I0309 18:49:04.212907 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerStarted","Data":"58a88d0572f6c186af4679d4d704bf3488b258adc08952d17236d4ce8fdcabf2"} Mar 09 18:49:05 crc kubenswrapper[4750]: I0309 18:49:05.227472 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerStarted","Data":"3d0bdb1caf11e5bdf51ac91b4902bdb495048802ee1f32e4c78d433101dc8612"} Mar 09 18:49:05 crc kubenswrapper[4750]: I0309 18:49:05.228816 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerStarted","Data":"962532b58e1a67163f363f4a9f7a6c9bb83b6aecc2ec29253029c99baa1b7b13"} Mar 09 18:49:05 crc kubenswrapper[4750]: I0309 18:49:05.611553 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:49:05 crc kubenswrapper[4750]: I0309 18:49:05.612526 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-log" containerID="cri-o://6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670" gracePeriod=30 Mar 09 18:49:05 crc kubenswrapper[4750]: I0309 18:49:05.613347 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-httpd" containerID="cri-o://976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1" gracePeriod=30 Mar 09 18:49:06 crc kubenswrapper[4750]: I0309 18:49:06.238972 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerStarted","Data":"31dc9deee4ff02237eab833925629b701205b732e144f18aa0501e9326f71dd6"} Mar 09 18:49:06 crc kubenswrapper[4750]: I0309 18:49:06.241383 4750 generic.go:334] "Generic (PLEG): container finished" podID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerID="6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670" exitCode=143 Mar 09 18:49:06 crc kubenswrapper[4750]: I0309 18:49:06.241452 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d177b683-bf5e-4048-a0cf-2b7e6176f2ed","Type":"ContainerDied","Data":"6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670"} Mar 09 18:49:07 crc kubenswrapper[4750]: I0309 18:49:07.082349 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:49:07 crc kubenswrapper[4750]: I0309 18:49:07.082671 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-log" containerID="cri-o://f4f3af90d447b2735d6453a83cbcee8ac25b9534a12865d87df3bc2bde9048a6" gracePeriod=30 Mar 09 18:49:07 crc kubenswrapper[4750]: I0309 18:49:07.082752 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-httpd" containerID="cri-o://2a713089dd3eef866ef549e6aae03e6e727a581f21b374c170ba07ed6358385b" gracePeriod=30 Mar 09 18:49:07 crc kubenswrapper[4750]: I0309 18:49:07.261395 4750 generic.go:334] "Generic (PLEG): container finished" podID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerID="f4f3af90d447b2735d6453a83cbcee8ac25b9534a12865d87df3bc2bde9048a6" exitCode=143 Mar 09 18:49:07 crc kubenswrapper[4750]: I0309 18:49:07.261822 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e","Type":"ContainerDied","Data":"f4f3af90d447b2735d6453a83cbcee8ac25b9534a12865d87df3bc2bde9048a6"} Mar 09 18:49:07 crc kubenswrapper[4750]: I0309 18:49:07.949087 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.146893 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-scripts\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.146986 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.147161 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-httpd-run\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.147209 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-combined-ca-bundle\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.147249 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-public-tls-certs\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.147272 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-config-data\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.147321 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-logs\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.147344 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j65hg\" (UniqueName: \"kubernetes.io/projected/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-kube-api-access-j65hg\") pod \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\" (UID: \"d177b683-bf5e-4048-a0cf-2b7e6176f2ed\") " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.148605 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-logs" (OuterVolumeSpecName: "logs") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.148869 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.154185 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-scripts" (OuterVolumeSpecName: "scripts") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.156067 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-kube-api-access-j65hg" (OuterVolumeSpecName: "kube-api-access-j65hg") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "kube-api-access-j65hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.158556 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.182701 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.213357 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-config-data" (OuterVolumeSpecName: "config-data") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.226113 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d177b683-bf5e-4048-a0cf-2b7e6176f2ed" (UID: "d177b683-bf5e-4048-a0cf-2b7e6176f2ed"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249714 4750 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249748 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249759 4750 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249768 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249778 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249788 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j65hg\" (UniqueName: \"kubernetes.io/projected/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-kube-api-access-j65hg\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249796 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d177b683-bf5e-4048-a0cf-2b7e6176f2ed-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.249832 4750 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.272506 4750 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.286261 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerStarted","Data":"40073f19e29398725de20d0df1e18c2d4bc69c28c7c407b45909934bd363077b"} Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.286348 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.291687 4750 generic.go:334] "Generic (PLEG): container finished" podID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerID="976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1" exitCode=0 Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.291743 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d177b683-bf5e-4048-a0cf-2b7e6176f2ed","Type":"ContainerDied","Data":"976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1"} Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.291777 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d177b683-bf5e-4048-a0cf-2b7e6176f2ed","Type":"ContainerDied","Data":"6a2c6af7749343b6d46928b0042c01d49f80d01d14b2d99b0c02e006736f62c0"} Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.291800 4750 scope.go:117] "RemoveContainer" containerID="976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.291984 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.312411 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.132339622 podStartE2EDuration="5.312389725s" podCreationTimestamp="2026-03-09 18:49:03 +0000 UTC" firstStartedPulling="2026-03-09 18:49:04.123212828 +0000 UTC m=+1425.465685226" lastFinishedPulling="2026-03-09 18:49:07.303262921 +0000 UTC m=+1428.645735329" observedRunningTime="2026-03-09 18:49:08.308647804 +0000 UTC m=+1429.651120202" watchObservedRunningTime="2026-03-09 18:49:08.312389725 +0000 UTC m=+1429.654862123" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.329901 4750 scope.go:117] "RemoveContainer" containerID="6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.342781 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.352164 4750 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.353551 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.355212 4750 scope.go:117] "RemoveContainer" containerID="976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1" Mar 09 18:49:08 crc kubenswrapper[4750]: E0309 18:49:08.357158 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1\": container with ID starting with 976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1 not found: ID does not exist" containerID="976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.357205 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1"} err="failed to get container status \"976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1\": rpc error: code = NotFound desc = could not find container \"976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1\": container with ID starting with 976fc747d892792ae721ea9822f6003a82fa13647f712f1ec2226bcac7f568a1 not found: ID does not exist" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.357242 4750 scope.go:117] "RemoveContainer" containerID="6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670" Mar 09 18:49:08 crc kubenswrapper[4750]: E0309 18:49:08.364549 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670\": container with ID starting with 6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670 not found: ID does not exist" containerID="6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.364610 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670"} err="failed to get container status \"6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670\": rpc error: code = NotFound desc = could not find container \"6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670\": container with ID starting with 6294e34b50ea752c62f4f319c590bac94472592bf3b4b84c4b092f378aa47670 not found: ID does not exist" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.378321 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:49:08 crc kubenswrapper[4750]: E0309 18:49:08.379043 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-log" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.379065 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-log" Mar 09 18:49:08 crc kubenswrapper[4750]: E0309 18:49:08.379103 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-httpd" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.379115 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-httpd" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.379427 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-httpd" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.379462 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" containerName="glance-log" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.381091 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.383868 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.384019 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.391482 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.558940 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.559003 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-logs\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.559037 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.559198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.559294 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfk7m\" (UniqueName: \"kubernetes.io/projected/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-kube-api-access-kfk7m\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.559388 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.559483 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.559788 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661543 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661589 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-logs\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661621 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661705 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661739 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfk7m\" (UniqueName: \"kubernetes.io/projected/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-kube-api-access-kfk7m\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661772 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661812 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.661847 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.662417 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.662676 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-logs\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.663134 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.665980 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.666170 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.670107 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.681205 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.689895 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfk7m\" (UniqueName: \"kubernetes.io/projected/3e04ac5d-19d1-4a07-846e-22852e6a0ca6-kube-api-access-kfk7m\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.699979 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3e04ac5d-19d1-4a07-846e-22852e6a0ca6\") " pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.712657 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 09 18:49:08 crc kubenswrapper[4750]: I0309 18:49:08.989226 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.028195 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.372920 4750 generic.go:334] "Generic (PLEG): container finished" podID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerID="2a713089dd3eef866ef549e6aae03e6e727a581f21b374c170ba07ed6358385b" exitCode=0 Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.373068 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e","Type":"ContainerDied","Data":"2a713089dd3eef866ef549e6aae03e6e727a581f21b374c170ba07ed6358385b"} Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.431195 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d177b683-bf5e-4048-a0cf-2b7e6176f2ed" path="/var/lib/kubelet/pods/d177b683-bf5e-4048-a0cf-2b7e6176f2ed/volumes" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.432096 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.468890 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.497484 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.525395 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.543856 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.602745 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.602785 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-combined-ca-bundle\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.602854 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-config-data\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.602890 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-internal-tls-certs\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.602941 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-scripts\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.603008 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-logs\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.603058 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-httpd-run\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.603260 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t87kn\" (UniqueName: \"kubernetes.io/projected/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-kube-api-access-t87kn\") pod \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\" (UID: \"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e\") " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.623152 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-logs" (OuterVolumeSpecName: "logs") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.626449 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.660132 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-scripts" (OuterVolumeSpecName: "scripts") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.660275 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-kube-api-access-t87kn" (OuterVolumeSpecName: "kube-api-access-t87kn") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "kube-api-access-t87kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.660364 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.706564 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.706598 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.706607 4750 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.706642 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t87kn\" (UniqueName: \"kubernetes.io/projected/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-kube-api-access-t87kn\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.706667 4750 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.720546 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.740416 4750 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.744878 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.751536 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-config-data" (OuterVolumeSpecName: "config-data") pod "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" (UID: "91ab0cb5-95bc-4e78-9acc-9a0414c6c25e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.808577 4750 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.808626 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.809186 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:09 crc kubenswrapper[4750]: I0309 18:49:09.809199 4750 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.353059 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.401882 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91ab0cb5-95bc-4e78-9acc-9a0414c6c25e","Type":"ContainerDied","Data":"d618d2e49e51c911395fe859fc17574fe7574c814eb40f6e5cb5962957e8062e"} Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.402455 4750 scope.go:117] "RemoveContainer" containerID="2a713089dd3eef866ef549e6aae03e6e727a581f21b374c170ba07ed6358385b" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.402674 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.412908 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e04ac5d-19d1-4a07-846e-22852e6a0ca6","Type":"ContainerStarted","Data":"4ba1d3d733b802792c914ff1e856211640dc2b4da8921698719e38492c189695"} Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.412942 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e04ac5d-19d1-4a07-846e-22852e6a0ca6","Type":"ContainerStarted","Data":"af3470469e26e35d346426f22715ba94856382f014af405a250f765bad7a361f"} Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.413150 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-central-agent" containerID="cri-o://962532b58e1a67163f363f4a9f7a6c9bb83b6aecc2ec29253029c99baa1b7b13" gracePeriod=30 Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.413856 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="proxy-httpd" containerID="cri-o://40073f19e29398725de20d0df1e18c2d4bc69c28c7c407b45909934bd363077b" gracePeriod=30 Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.413941 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="sg-core" containerID="cri-o://31dc9deee4ff02237eab833925629b701205b732e144f18aa0501e9326f71dd6" gracePeriod=30 Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.413998 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-notification-agent" containerID="cri-o://3d0bdb1caf11e5bdf51ac91b4902bdb495048802ee1f32e4c78d433101dc8612" gracePeriod=30 Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.483830 4750 scope.go:117] "RemoveContainer" containerID="f4f3af90d447b2735d6453a83cbcee8ac25b9534a12865d87df3bc2bde9048a6" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.488774 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.497559 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.510844 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:49:10 crc kubenswrapper[4750]: E0309 18:49:10.511274 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-httpd" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.511293 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-httpd" Mar 09 18:49:10 crc kubenswrapper[4750]: E0309 18:49:10.511318 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-log" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.511325 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-log" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.511536 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-httpd" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.511560 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" containerName="glance-log" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.512620 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.522200 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.522417 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.528839 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.643478 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.643624 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfm4g\" (UniqueName: \"kubernetes.io/projected/8b839eed-70cc-4143-880a-52591ffb61c5-kube-api-access-wfm4g\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.643703 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.643732 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.643803 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b839eed-70cc-4143-880a-52591ffb61c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.643833 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.643891 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.644073 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b839eed-70cc-4143-880a-52591ffb61c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.749962 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750044 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfm4g\" (UniqueName: \"kubernetes.io/projected/8b839eed-70cc-4143-880a-52591ffb61c5-kube-api-access-wfm4g\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750080 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750111 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750138 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b839eed-70cc-4143-880a-52591ffb61c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750171 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750201 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750240 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b839eed-70cc-4143-880a-52591ffb61c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.750907 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b839eed-70cc-4143-880a-52591ffb61c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.751122 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b839eed-70cc-4143-880a-52591ffb61c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.751765 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.758466 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.760871 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.780427 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.782525 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfm4g\" (UniqueName: \"kubernetes.io/projected/8b839eed-70cc-4143-880a-52591ffb61c5-kube-api-access-wfm4g\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.784889 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b839eed-70cc-4143-880a-52591ffb61c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.840278 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b839eed-70cc-4143-880a-52591ffb61c5\") " pod="openstack/glance-default-internal-api-0" Mar 09 18:49:10 crc kubenswrapper[4750]: I0309 18:49:10.886799 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.210698 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.211546 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="c857b2a0-c8c1-452f-99af-a6affa05c8c1" containerName="watcher-applier" containerID="cri-o://36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74" gracePeriod=30 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.248283 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.248595 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api-log" containerID="cri-o://81fa205da7414242348162b76897b36b68057304538aeb66e5fce95f2da52539" gracePeriod=30 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.250882 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api" containerID="cri-o://31ac876a95bf09d01ad8b09379fa6625140fab0337498dd49db09774a709a665" gracePeriod=30 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.387861 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91ab0cb5-95bc-4e78-9acc-9a0414c6c25e" path="/var/lib/kubelet/pods/91ab0cb5-95bc-4e78-9acc-9a0414c6c25e/volumes" Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.433292 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e04ac5d-19d1-4a07-846e-22852e6a0ca6","Type":"ContainerStarted","Data":"b96ac85fc4c5c752149af5edc1c52057c0a256e76b9d525b215055473487c505"} Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.446076 4750 generic.go:334] "Generic (PLEG): container finished" podID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerID="81fa205da7414242348162b76897b36b68057304538aeb66e5fce95f2da52539" exitCode=143 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.446169 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e8060c58-bb40-45af-86d4-310cafdcfde6","Type":"ContainerDied","Data":"81fa205da7414242348162b76897b36b68057304538aeb66e5fce95f2da52539"} Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.455735 4750 generic.go:334] "Generic (PLEG): container finished" podID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerID="40073f19e29398725de20d0df1e18c2d4bc69c28c7c407b45909934bd363077b" exitCode=0 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.455770 4750 generic.go:334] "Generic (PLEG): container finished" podID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerID="31dc9deee4ff02237eab833925629b701205b732e144f18aa0501e9326f71dd6" exitCode=2 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.455779 4750 generic.go:334] "Generic (PLEG): container finished" podID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerID="3d0bdb1caf11e5bdf51ac91b4902bdb495048802ee1f32e4c78d433101dc8612" exitCode=0 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.456056 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" containerID="cri-o://e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95" gracePeriod=30 Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.456325 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerDied","Data":"40073f19e29398725de20d0df1e18c2d4bc69c28c7c407b45909934bd363077b"} Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.456359 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerDied","Data":"31dc9deee4ff02237eab833925629b701205b732e144f18aa0501e9326f71dd6"} Mar 09 18:49:11 crc kubenswrapper[4750]: E0309 18:49:11.456339 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8060c58_bb40_45af_86d4_310cafdcfde6.slice/crio-81fa205da7414242348162b76897b36b68057304538aeb66e5fce95f2da52539.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8060c58_bb40_45af_86d4_310cafdcfde6.slice/crio-conmon-81fa205da7414242348162b76897b36b68057304538aeb66e5fce95f2da52539.scope\": RecentStats: unable to find data in memory cache]" Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.456370 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerDied","Data":"3d0bdb1caf11e5bdf51ac91b4902bdb495048802ee1f32e4c78d433101dc8612"} Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.468158 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.468134988 podStartE2EDuration="3.468134988s" podCreationTimestamp="2026-03-09 18:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:49:11.460272714 +0000 UTC m=+1432.802745112" watchObservedRunningTime="2026-03-09 18:49:11.468134988 +0000 UTC m=+1432.810607386" Mar 09 18:49:11 crc kubenswrapper[4750]: I0309 18:49:11.513581 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 09 18:49:12 crc kubenswrapper[4750]: E0309 18:49:12.333174 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 09 18:49:12 crc kubenswrapper[4750]: E0309 18:49:12.350826 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 09 18:49:12 crc kubenswrapper[4750]: E0309 18:49:12.376679 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 09 18:49:12 crc kubenswrapper[4750]: E0309 18:49:12.376754 4750 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="c857b2a0-c8c1-452f-99af-a6affa05c8c1" containerName="watcher-applier" Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.529885 4750 generic.go:334] "Generic (PLEG): container finished" podID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerID="31ac876a95bf09d01ad8b09379fa6625140fab0337498dd49db09774a709a665" exitCode=0 Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.530005 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e8060c58-bb40-45af-86d4-310cafdcfde6","Type":"ContainerDied","Data":"31ac876a95bf09d01ad8b09379fa6625140fab0337498dd49db09774a709a665"} Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.606754 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8b839eed-70cc-4143-880a-52591ffb61c5","Type":"ContainerStarted","Data":"b703dd4a18424446eb403c5990c70cd7b0be7aa9f3205c47ee4838d39bd3eb6d"} Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.607348 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8b839eed-70cc-4143-880a-52591ffb61c5","Type":"ContainerStarted","Data":"207f59f45bdb01afafcf80d12a033bad0a636663fe530749bb72fcb2ed5ab988"} Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.733813 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.928433 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-custom-prometheus-ca\") pod \"e8060c58-bb40-45af-86d4-310cafdcfde6\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.928486 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-public-tls-certs\") pod \"e8060c58-bb40-45af-86d4-310cafdcfde6\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.928576 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-config-data\") pod \"e8060c58-bb40-45af-86d4-310cafdcfde6\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.928596 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-combined-ca-bundle\") pod \"e8060c58-bb40-45af-86d4-310cafdcfde6\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.928705 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-internal-tls-certs\") pod \"e8060c58-bb40-45af-86d4-310cafdcfde6\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.928762 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8060c58-bb40-45af-86d4-310cafdcfde6-logs\") pod \"e8060c58-bb40-45af-86d4-310cafdcfde6\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.929256 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8060c58-bb40-45af-86d4-310cafdcfde6-logs" (OuterVolumeSpecName: "logs") pod "e8060c58-bb40-45af-86d4-310cafdcfde6" (UID: "e8060c58-bb40-45af-86d4-310cafdcfde6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.929340 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6jxq\" (UniqueName: \"kubernetes.io/projected/e8060c58-bb40-45af-86d4-310cafdcfde6-kube-api-access-x6jxq\") pod \"e8060c58-bb40-45af-86d4-310cafdcfde6\" (UID: \"e8060c58-bb40-45af-86d4-310cafdcfde6\") " Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.930776 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8060c58-bb40-45af-86d4-310cafdcfde6-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.942531 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8060c58-bb40-45af-86d4-310cafdcfde6-kube-api-access-x6jxq" (OuterVolumeSpecName: "kube-api-access-x6jxq") pod "e8060c58-bb40-45af-86d4-310cafdcfde6" (UID: "e8060c58-bb40-45af-86d4-310cafdcfde6"). InnerVolumeSpecName "kube-api-access-x6jxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.974953 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "e8060c58-bb40-45af-86d4-310cafdcfde6" (UID: "e8060c58-bb40-45af-86d4-310cafdcfde6"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.975790 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8060c58-bb40-45af-86d4-310cafdcfde6" (UID: "e8060c58-bb40-45af-86d4-310cafdcfde6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:12 crc kubenswrapper[4750]: I0309 18:49:12.985334 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e8060c58-bb40-45af-86d4-310cafdcfde6" (UID: "e8060c58-bb40-45af-86d4-310cafdcfde6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.020167 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e8060c58-bb40-45af-86d4-310cafdcfde6" (UID: "e8060c58-bb40-45af-86d4-310cafdcfde6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.025753 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-config-data" (OuterVolumeSpecName: "config-data") pod "e8060c58-bb40-45af-86d4-310cafdcfde6" (UID: "e8060c58-bb40-45af-86d4-310cafdcfde6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.033583 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6jxq\" (UniqueName: \"kubernetes.io/projected/e8060c58-bb40-45af-86d4-310cafdcfde6-kube-api-access-x6jxq\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.033618 4750 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.033650 4750 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.033664 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.033673 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.033682 4750 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8060c58-bb40-45af-86d4-310cafdcfde6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.619348 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e8060c58-bb40-45af-86d4-310cafdcfde6","Type":"ContainerDied","Data":"16c80cf1aa66d72a714d62df4fbecd17f022e8e61530e69c0086a3a396f3a207"} Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.619446 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.619709 4750 scope.go:117] "RemoveContainer" containerID="31ac876a95bf09d01ad8b09379fa6625140fab0337498dd49db09774a709a665" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.623178 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8b839eed-70cc-4143-880a-52591ffb61c5","Type":"ContainerStarted","Data":"6d911cf53002032fe3c46e69aa289e73b29fe162415793c9e4260a9ff69af7f8"} Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.648249 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.648232406 podStartE2EDuration="3.648232406s" podCreationTimestamp="2026-03-09 18:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:49:13.648161665 +0000 UTC m=+1434.990634073" watchObservedRunningTime="2026-03-09 18:49:13.648232406 +0000 UTC m=+1434.990704794" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.663037 4750 scope.go:117] "RemoveContainer" containerID="81fa205da7414242348162b76897b36b68057304538aeb66e5fce95f2da52539" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.682217 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.695956 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.711561 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:49:13 crc kubenswrapper[4750]: E0309 18:49:13.712064 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api-log" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.712094 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api-log" Mar 09 18:49:13 crc kubenswrapper[4750]: E0309 18:49:13.712120 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.712126 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.712337 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api-log" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.712361 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" containerName="watcher-api" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.713490 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.721917 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.726181 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.726872 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.727042 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.850734 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.850885 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.851011 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664aaf56-bbea-4a12-95d5-ab56a978d608-logs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.851509 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-config-data\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.851800 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-public-tls-certs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.851966 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slqn5\" (UniqueName: \"kubernetes.io/projected/664aaf56-bbea-4a12-95d5-ab56a978d608-kube-api-access-slqn5\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.852008 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.955615 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.955738 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.955794 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664aaf56-bbea-4a12-95d5-ab56a978d608-logs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.955848 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-config-data\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.955914 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-public-tls-certs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.955966 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slqn5\" (UniqueName: \"kubernetes.io/projected/664aaf56-bbea-4a12-95d5-ab56a978d608-kube-api-access-slqn5\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.955992 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.960002 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664aaf56-bbea-4a12-95d5-ab56a978d608-logs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.963239 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.963721 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.965736 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-config-data\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.965951 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-public-tls-certs\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.980942 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/664aaf56-bbea-4a12-95d5-ab56a978d608-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:13 crc kubenswrapper[4750]: I0309 18:49:13.990207 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slqn5\" (UniqueName: \"kubernetes.io/projected/664aaf56-bbea-4a12-95d5-ab56a978d608-kube-api-access-slqn5\") pod \"watcher-api-0\" (UID: \"664aaf56-bbea-4a12-95d5-ab56a978d608\") " pod="openstack/watcher-api-0" Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.056973 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.579272 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.643701 4750 generic.go:334] "Generic (PLEG): container finished" podID="c857b2a0-c8c1-452f-99af-a6affa05c8c1" containerID="36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74" exitCode=0 Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.643807 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"c857b2a0-c8c1-452f-99af-a6affa05c8c1","Type":"ContainerDied","Data":"36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74"} Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.646609 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"664aaf56-bbea-4a12-95d5-ab56a978d608","Type":"ContainerStarted","Data":"affb26d72827ab0725fb79e65ace61aad43fb51a47d26ae31ce28e7f04d0c3eb"} Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.885199 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.919760 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-combined-ca-bundle\") pod \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.920276 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9mwh\" (UniqueName: \"kubernetes.io/projected/c857b2a0-c8c1-452f-99af-a6affa05c8c1-kube-api-access-m9mwh\") pod \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.920307 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-config-data\") pod \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.920330 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c857b2a0-c8c1-452f-99af-a6affa05c8c1-logs\") pod \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\" (UID: \"c857b2a0-c8c1-452f-99af-a6affa05c8c1\") " Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.921548 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c857b2a0-c8c1-452f-99af-a6affa05c8c1-logs" (OuterVolumeSpecName: "logs") pod "c857b2a0-c8c1-452f-99af-a6affa05c8c1" (UID: "c857b2a0-c8c1-452f-99af-a6affa05c8c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.929769 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c857b2a0-c8c1-452f-99af-a6affa05c8c1-kube-api-access-m9mwh" (OuterVolumeSpecName: "kube-api-access-m9mwh") pod "c857b2a0-c8c1-452f-99af-a6affa05c8c1" (UID: "c857b2a0-c8c1-452f-99af-a6affa05c8c1"). InnerVolumeSpecName "kube-api-access-m9mwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:14 crc kubenswrapper[4750]: I0309 18:49:14.956892 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c857b2a0-c8c1-452f-99af-a6affa05c8c1" (UID: "c857b2a0-c8c1-452f-99af-a6affa05c8c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.022717 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.022752 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9mwh\" (UniqueName: \"kubernetes.io/projected/c857b2a0-c8c1-452f-99af-a6affa05c8c1-kube-api-access-m9mwh\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.022765 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c857b2a0-c8c1-452f-99af-a6affa05c8c1-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.022748 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-config-data" (OuterVolumeSpecName: "config-data") pod "c857b2a0-c8c1-452f-99af-a6affa05c8c1" (UID: "c857b2a0-c8c1-452f-99af-a6affa05c8c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.125106 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c857b2a0-c8c1-452f-99af-a6affa05c8c1-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.385524 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8060c58-bb40-45af-86d4-310cafdcfde6" path="/var/lib/kubelet/pods/e8060c58-bb40-45af-86d4-310cafdcfde6/volumes" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.657191 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"664aaf56-bbea-4a12-95d5-ab56a978d608","Type":"ContainerStarted","Data":"bf60d97c34498121f4e90ef464b96c557bd92e850e4fa16b0f2b04488029d012"} Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.658430 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"664aaf56-bbea-4a12-95d5-ab56a978d608","Type":"ContainerStarted","Data":"f2709ee1247b341dd929fb3e378eeb6ff748fb1857c8b4b8aaba5c2788cfa855"} Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.658525 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.660188 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"c857b2a0-c8c1-452f-99af-a6affa05c8c1","Type":"ContainerDied","Data":"424ff5429ce057d88b79873fda82e75df276b23d0d0646f94f9628df631e8d9c"} Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.660318 4750 scope.go:117] "RemoveContainer" containerID="36135d8636957dc13b1b978c54f0c0efaff5fa956472ec878a8680e3ebbd8e74" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.660526 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.689695 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=2.689677929 podStartE2EDuration="2.689677929s" podCreationTimestamp="2026-03-09 18:49:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:49:15.686243016 +0000 UTC m=+1437.028715434" watchObservedRunningTime="2026-03-09 18:49:15.689677929 +0000 UTC m=+1437.032150327" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.711895 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.720837 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.747523 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:49:15 crc kubenswrapper[4750]: E0309 18:49:15.750004 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c857b2a0-c8c1-452f-99af-a6affa05c8c1" containerName="watcher-applier" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.750033 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c857b2a0-c8c1-452f-99af-a6affa05c8c1" containerName="watcher-applier" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.750218 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c857b2a0-c8c1-452f-99af-a6affa05c8c1" containerName="watcher-applier" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.750951 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.754037 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.781269 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.838178 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt94f\" (UniqueName: \"kubernetes.io/projected/d7f9e82f-cb7a-425d-b942-b41d9c167784-kube-api-access-xt94f\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.838271 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7f9e82f-cb7a-425d-b942-b41d9c167784-config-data\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.838369 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7f9e82f-cb7a-425d-b942-b41d9c167784-logs\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.838394 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f9e82f-cb7a-425d-b942-b41d9c167784-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.941607 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7f9e82f-cb7a-425d-b942-b41d9c167784-logs\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.941664 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f9e82f-cb7a-425d-b942-b41d9c167784-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.941779 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt94f\" (UniqueName: \"kubernetes.io/projected/d7f9e82f-cb7a-425d-b942-b41d9c167784-kube-api-access-xt94f\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.941814 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7f9e82f-cb7a-425d-b942-b41d9c167784-config-data\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.942229 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7f9e82f-cb7a-425d-b942-b41d9c167784-logs\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.950696 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7f9e82f-cb7a-425d-b942-b41d9c167784-config-data\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.950739 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f9e82f-cb7a-425d-b942-b41d9c167784-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:15 crc kubenswrapper[4750]: I0309 18:49:15.989536 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt94f\" (UniqueName: \"kubernetes.io/projected/d7f9e82f-cb7a-425d-b942-b41d9c167784-kube-api-access-xt94f\") pod \"watcher-applier-0\" (UID: \"d7f9e82f-cb7a-425d-b942-b41d9c167784\") " pod="openstack/watcher-applier-0" Mar 09 18:49:16 crc kubenswrapper[4750]: I0309 18:49:16.069943 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Mar 09 18:49:16 crc kubenswrapper[4750]: I0309 18:49:16.590993 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Mar 09 18:49:16 crc kubenswrapper[4750]: I0309 18:49:16.684128 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"d7f9e82f-cb7a-425d-b942-b41d9c167784","Type":"ContainerStarted","Data":"5f44c0fd1594c415c6bca9620bde2f3f480fa7c3a87160bd312cffef9fb1c409"} Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.163306 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.271206 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6796c4d9-5680-4c22-9ef0-815642b22c3f-logs\") pod \"6796c4d9-5680-4c22-9ef0-815642b22c3f\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.271280 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgqsm\" (UniqueName: \"kubernetes.io/projected/6796c4d9-5680-4c22-9ef0-815642b22c3f-kube-api-access-zgqsm\") pod \"6796c4d9-5680-4c22-9ef0-815642b22c3f\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.271325 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-custom-prometheus-ca\") pod \"6796c4d9-5680-4c22-9ef0-815642b22c3f\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.271349 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-combined-ca-bundle\") pod \"6796c4d9-5680-4c22-9ef0-815642b22c3f\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.271430 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-config-data\") pod \"6796c4d9-5680-4c22-9ef0-815642b22c3f\" (UID: \"6796c4d9-5680-4c22-9ef0-815642b22c3f\") " Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.271583 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6796c4d9-5680-4c22-9ef0-815642b22c3f-logs" (OuterVolumeSpecName: "logs") pod "6796c4d9-5680-4c22-9ef0-815642b22c3f" (UID: "6796c4d9-5680-4c22-9ef0-815642b22c3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.272424 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6796c4d9-5680-4c22-9ef0-815642b22c3f-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.277079 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6796c4d9-5680-4c22-9ef0-815642b22c3f-kube-api-access-zgqsm" (OuterVolumeSpecName: "kube-api-access-zgqsm") pod "6796c4d9-5680-4c22-9ef0-815642b22c3f" (UID: "6796c4d9-5680-4c22-9ef0-815642b22c3f"). InnerVolumeSpecName "kube-api-access-zgqsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.304706 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "6796c4d9-5680-4c22-9ef0-815642b22c3f" (UID: "6796c4d9-5680-4c22-9ef0-815642b22c3f"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.308240 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6796c4d9-5680-4c22-9ef0-815642b22c3f" (UID: "6796c4d9-5680-4c22-9ef0-815642b22c3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.348240 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-config-data" (OuterVolumeSpecName: "config-data") pod "6796c4d9-5680-4c22-9ef0-815642b22c3f" (UID: "6796c4d9-5680-4c22-9ef0-815642b22c3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.380807 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgqsm\" (UniqueName: \"kubernetes.io/projected/6796c4d9-5680-4c22-9ef0-815642b22c3f-kube-api-access-zgqsm\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.381095 4750 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.381165 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.381226 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6796c4d9-5680-4c22-9ef0-815642b22c3f-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.418333 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c857b2a0-c8c1-452f-99af-a6affa05c8c1" path="/var/lib/kubelet/pods/c857b2a0-c8c1-452f-99af-a6affa05c8c1/volumes" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.708866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"d7f9e82f-cb7a-425d-b942-b41d9c167784","Type":"ContainerStarted","Data":"cc0dd3950c72c4bcd0db02ad10433a5f2df13d155bb15dc82e5dfcb83e3ecf27"} Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.712714 4750 generic.go:334] "Generic (PLEG): container finished" podID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerID="e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95" exitCode=0 Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.712781 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerDied","Data":"e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95"} Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.712814 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"6796c4d9-5680-4c22-9ef0-815642b22c3f","Type":"ContainerDied","Data":"bc6a490fec8c24319e9b4a56ade1c502d219edbe8c0c2af72a71fc412d4a4c9d"} Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.712835 4750 scope.go:117] "RemoveContainer" containerID="e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.713897 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.737408 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=2.7373824620000002 podStartE2EDuration="2.737382462s" podCreationTimestamp="2026-03-09 18:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:49:17.726255549 +0000 UTC m=+1439.068727957" watchObservedRunningTime="2026-03-09 18:49:17.737382462 +0000 UTC m=+1439.079854860" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.763810 4750 scope.go:117] "RemoveContainer" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.773759 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.790739 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.832724 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:49:17 crc kubenswrapper[4750]: E0309 18:49:17.833191 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.833203 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: E0309 18:49:17.833224 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.833230 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: E0309 18:49:17.833238 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.833245 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.833426 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.833439 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.834123 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.840572 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.847953 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.890378 4750 scope.go:117] "RemoveContainer" containerID="e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95" Mar 09 18:49:17 crc kubenswrapper[4750]: E0309 18:49:17.891286 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95\": container with ID starting with e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95 not found: ID does not exist" containerID="e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.891341 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95"} err="failed to get container status \"e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95\": rpc error: code = NotFound desc = could not find container \"e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95\": container with ID starting with e8fe3894d22f33e093c2a053892175ca6e3232c607b87a291a16524379a69a95 not found: ID does not exist" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.891373 4750 scope.go:117] "RemoveContainer" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.891874 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.891995 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-config-data\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.892100 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-logs\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.892233 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvzv4\" (UniqueName: \"kubernetes.io/projected/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-kube-api-access-fvzv4\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.892364 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: E0309 18:49:17.894467 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f\": container with ID starting with eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f not found: ID does not exist" containerID="eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.894512 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f"} err="failed to get container status \"eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f\": rpc error: code = NotFound desc = could not find container \"eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f\": container with ID starting with eae16309319130e5c93036c34fcd8d17dfe6e3b806f317fb0623718e63f86e9f not found: ID does not exist" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.994476 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.994543 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-config-data\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.994561 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-logs\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.994618 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvzv4\" (UniqueName: \"kubernetes.io/projected/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-kube-api-access-fvzv4\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.994713 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.995547 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-logs\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:17 crc kubenswrapper[4750]: I0309 18:49:17.999527 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.000176 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-config-data\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.000228 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.014039 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvzv4\" (UniqueName: \"kubernetes.io/projected/ab65e149-9e3c-451f-82a4-e0fbe5127ac1-kube-api-access-fvzv4\") pod \"watcher-decision-engine-0\" (UID: \"ab65e149-9e3c-451f-82a4-e0fbe5127ac1\") " pod="openstack/watcher-decision-engine-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.194524 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.513233 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.652419 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.713314 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.713368 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.726035 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"ab65e149-9e3c-451f-82a4-e0fbe5127ac1","Type":"ContainerStarted","Data":"19ef3c99447de930898544242dc00b9b35ece764250b60c491309fc36d2b23ab"} Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.767492 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 09 18:49:18 crc kubenswrapper[4750]: I0309 18:49:18.767571 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.057730 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.391457 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" path="/var/lib/kubelet/pods/6796c4d9-5680-4c22-9ef0-815642b22c3f/volumes" Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.742034 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"ab65e149-9e3c-451f-82a4-e0fbe5127ac1","Type":"ContainerStarted","Data":"529bb27d81e1e5a63ff70c65236f9ecbc993bcf144de89ba324777b209f3b009"} Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.744223 4750 generic.go:334] "Generic (PLEG): container finished" podID="31e486b2-f819-4c3a-b7db-2262ad76bbb2" containerID="7764ae635f169cc2206c2082440203a3b76a001e64c8eab35587e01281ad3419" exitCode=0 Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.744314 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" event={"ID":"31e486b2-f819-4c3a-b7db-2262ad76bbb2","Type":"ContainerDied","Data":"7764ae635f169cc2206c2082440203a3b76a001e64c8eab35587e01281ad3419"} Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.744647 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.744708 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 09 18:49:19 crc kubenswrapper[4750]: I0309 18:49:19.765761 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.765744367 podStartE2EDuration="2.765744367s" podCreationTimestamp="2026-03-09 18:49:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:49:19.764331639 +0000 UTC m=+1441.106804037" watchObservedRunningTime="2026-03-09 18:49:19.765744367 +0000 UTC m=+1441.108216765" Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.759137 4750 generic.go:334] "Generic (PLEG): container finished" podID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerID="962532b58e1a67163f363f4a9f7a6c9bb83b6aecc2ec29253029c99baa1b7b13" exitCode=0 Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.760857 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerDied","Data":"962532b58e1a67163f363f4a9f7a6c9bb83b6aecc2ec29253029c99baa1b7b13"} Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.853210 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.886968 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.887022 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.938109 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.948193 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.964848 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-sg-core-conf-yaml\") pod \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.964922 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-scripts\") pod \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.964974 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-config-data\") pod \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.965008 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-combined-ca-bundle\") pod \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.965254 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-run-httpd\") pod \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.965296 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-log-httpd\") pod \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.965373 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m85bz\" (UniqueName: \"kubernetes.io/projected/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-kube-api-access-m85bz\") pod \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\" (UID: \"ced5718e-bed8-44d0-a3e3-be97c80dc1d5\") " Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.979616 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ced5718e-bed8-44d0-a3e3-be97c80dc1d5" (UID: "ced5718e-bed8-44d0-a3e3-be97c80dc1d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:20 crc kubenswrapper[4750]: I0309 18:49:20.980573 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ced5718e-bed8-44d0-a3e3-be97c80dc1d5" (UID: "ced5718e-bed8-44d0-a3e3-be97c80dc1d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:20.996805 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-scripts" (OuterVolumeSpecName: "scripts") pod "ced5718e-bed8-44d0-a3e3-be97c80dc1d5" (UID: "ced5718e-bed8-44d0-a3e3-be97c80dc1d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:20.997044 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-kube-api-access-m85bz" (OuterVolumeSpecName: "kube-api-access-m85bz") pod "ced5718e-bed8-44d0-a3e3-be97c80dc1d5" (UID: "ced5718e-bed8-44d0-a3e3-be97c80dc1d5"). InnerVolumeSpecName "kube-api-access-m85bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.034030 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ced5718e-bed8-44d0-a3e3-be97c80dc1d5" (UID: "ced5718e-bed8-44d0-a3e3-be97c80dc1d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.068286 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.068316 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.068325 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m85bz\" (UniqueName: \"kubernetes.io/projected/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-kube-api-access-m85bz\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.068336 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.068345 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.070394 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.079921 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ced5718e-bed8-44d0-a3e3-be97c80dc1d5" (UID: "ced5718e-bed8-44d0-a3e3-be97c80dc1d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.137776 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.159908 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-config-data" (OuterVolumeSpecName: "config-data") pod "ced5718e-bed8-44d0-a3e3-be97c80dc1d5" (UID: "ced5718e-bed8-44d0-a3e3-be97c80dc1d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.180123 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.180161 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced5718e-bed8-44d0-a3e3-be97c80dc1d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.282154 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-combined-ca-bundle\") pod \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.282240 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-scripts\") pod \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.282521 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh9ds\" (UniqueName: \"kubernetes.io/projected/31e486b2-f819-4c3a-b7db-2262ad76bbb2-kube-api-access-wh9ds\") pod \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.282659 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-config-data\") pod \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\" (UID: \"31e486b2-f819-4c3a-b7db-2262ad76bbb2\") " Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.285768 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-scripts" (OuterVolumeSpecName: "scripts") pod "31e486b2-f819-4c3a-b7db-2262ad76bbb2" (UID: "31e486b2-f819-4c3a-b7db-2262ad76bbb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.289466 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e486b2-f819-4c3a-b7db-2262ad76bbb2-kube-api-access-wh9ds" (OuterVolumeSpecName: "kube-api-access-wh9ds") pod "31e486b2-f819-4c3a-b7db-2262ad76bbb2" (UID: "31e486b2-f819-4c3a-b7db-2262ad76bbb2"). InnerVolumeSpecName "kube-api-access-wh9ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.318813 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-config-data" (OuterVolumeSpecName: "config-data") pod "31e486b2-f819-4c3a-b7db-2262ad76bbb2" (UID: "31e486b2-f819-4c3a-b7db-2262ad76bbb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.333985 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31e486b2-f819-4c3a-b7db-2262ad76bbb2" (UID: "31e486b2-f819-4c3a-b7db-2262ad76bbb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.385659 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh9ds\" (UniqueName: \"kubernetes.io/projected/31e486b2-f819-4c3a-b7db-2262ad76bbb2-kube-api-access-wh9ds\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.385698 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.385710 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.385718 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31e486b2-f819-4c3a-b7db-2262ad76bbb2-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.781930 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ced5718e-bed8-44d0-a3e3-be97c80dc1d5","Type":"ContainerDied","Data":"58a88d0572f6c186af4679d4d704bf3488b258adc08952d17236d4ce8fdcabf2"} Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.781995 4750 scope.go:117] "RemoveContainer" containerID="40073f19e29398725de20d0df1e18c2d4bc69c28c7c407b45909934bd363077b" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.782172 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.802605 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.802760 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5jn6" event={"ID":"31e486b2-f819-4c3a-b7db-2262ad76bbb2","Type":"ContainerDied","Data":"653f381241c4443d0373cddacc56767026eeffa91c73997c0081ba132d6875a2"} Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.802815 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="653f381241c4443d0373cddacc56767026eeffa91c73997c0081ba132d6875a2" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.803448 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.803566 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.830435 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.867745 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.879916 4750 scope.go:117] "RemoveContainer" containerID="31dc9deee4ff02237eab833925629b701205b732e144f18aa0501e9326f71dd6" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.896343 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:21 crc kubenswrapper[4750]: E0309 18:49:21.896798 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e486b2-f819-4c3a-b7db-2262ad76bbb2" containerName="nova-cell0-conductor-db-sync" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.896818 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e486b2-f819-4c3a-b7db-2262ad76bbb2" containerName="nova-cell0-conductor-db-sync" Mar 09 18:49:21 crc kubenswrapper[4750]: E0309 18:49:21.896829 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-notification-agent" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.896836 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-notification-agent" Mar 09 18:49:21 crc kubenswrapper[4750]: E0309 18:49:21.896844 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-central-agent" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.896851 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-central-agent" Mar 09 18:49:21 crc kubenswrapper[4750]: E0309 18:49:21.896862 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="sg-core" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.896869 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="sg-core" Mar 09 18:49:21 crc kubenswrapper[4750]: E0309 18:49:21.896896 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="proxy-httpd" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.896903 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="proxy-httpd" Mar 09 18:49:21 crc kubenswrapper[4750]: E0309 18:49:21.896923 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.896931 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.897137 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="proxy-httpd" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.897152 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.897166 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-notification-agent" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.897179 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="sg-core" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.897191 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e486b2-f819-4c3a-b7db-2262ad76bbb2" containerName="nova-cell0-conductor-db-sync" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.897206 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" containerName="ceilometer-central-agent" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.897722 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6796c4d9-5680-4c22-9ef0-815642b22c3f" containerName="watcher-decision-engine" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.914212 4750 scope.go:117] "RemoveContainer" containerID="3d0bdb1caf11e5bdf51ac91b4902bdb495048802ee1f32e4c78d433101dc8612" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.930180 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.931576 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.938494 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.939268 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.939432 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.939510 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.959395 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.977027 4750 scope.go:117] "RemoveContainer" containerID="962532b58e1a67163f363f4a9f7a6c9bb83b6aecc2ec29253029c99baa1b7b13" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.988029 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.992407 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 09 18:49:21 crc kubenswrapper[4750]: I0309 18:49:21.997989 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kljb7" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.006832 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-config-data\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.006938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.007008 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-scripts\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.007062 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4zzp\" (UniqueName: \"kubernetes.io/projected/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-kube-api-access-f4zzp\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.007112 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-run-httpd\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.007156 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.007241 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-log-httpd\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.033705 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.109787 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-config-data\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.109916 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.109974 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-scripts\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.110032 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4zzp\" (UniqueName: \"kubernetes.io/projected/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-kube-api-access-f4zzp\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.110118 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-run-httpd\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.110157 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.110192 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgmw7\" (UniqueName: \"kubernetes.io/projected/d2209408-2992-4bc3-b338-464fbaccf032-kube-api-access-wgmw7\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.110231 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.110291 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-log-httpd\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.110316 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.111474 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-run-httpd\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.111768 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-log-httpd\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.117440 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.117963 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-config-data\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.118180 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-scripts\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.119629 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.128451 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4zzp\" (UniqueName: \"kubernetes.io/projected/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-kube-api-access-f4zzp\") pod \"ceilometer-0\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.149473 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.212088 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.212307 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgmw7\" (UniqueName: \"kubernetes.io/projected/d2209408-2992-4bc3-b338-464fbaccf032-kube-api-access-wgmw7\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.212338 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.221495 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.221549 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.237437 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgmw7\" (UniqueName: \"kubernetes.io/projected/d2209408-2992-4bc3-b338-464fbaccf032-kube-api-access-wgmw7\") pod \"nova-cell0-conductor-0\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.277827 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.327312 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.795008 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:22 crc kubenswrapper[4750]: I0309 18:49:22.900814 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.393336 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced5718e-bed8-44d0-a3e3-be97c80dc1d5" path="/var/lib/kubelet/pods/ced5718e-bed8-44d0-a3e3-be97c80dc1d5/volumes" Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.768363 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.840217 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d2209408-2992-4bc3-b338-464fbaccf032","Type":"ContainerStarted","Data":"69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb"} Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.840282 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d2209408-2992-4bc3-b338-464fbaccf032","Type":"ContainerStarted","Data":"50917a2fd4c630bfcf6ab88531842a9197035648e327f9e54612a136167431a6"} Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.840383 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.842373 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.842394 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.842410 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerStarted","Data":"6d9ca2d1e938311fff282467d18b766a3e663e814778d34847a5a71b880d22c7"} Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.842466 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerStarted","Data":"d2577258b0ab86ef7614d4daf96710022ca9c392c23d14e6d0e2f620ea7fe61b"} Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.842483 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerStarted","Data":"5f36305c574a70eee9ac5a12003b5a4bb109346141b9e6f7f853226cae1518c2"} Mar 09 18:49:23 crc kubenswrapper[4750]: I0309 18:49:23.863682 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.862613723 podStartE2EDuration="2.862613723s" podCreationTimestamp="2026-03-09 18:49:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:49:23.855813548 +0000 UTC m=+1445.198285946" watchObservedRunningTime="2026-03-09 18:49:23.862613723 +0000 UTC m=+1445.205086121" Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.056789 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.057551 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.076583 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.447377 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.660938 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.855342 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerStarted","Data":"5a1c2c505559b89586bede407c9b6b585d862281994a05c348540b3f34a85029"} Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.861913 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" containerID="cri-o://69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" gracePeriod=30 Mar 09 18:49:24 crc kubenswrapper[4750]: I0309 18:49:24.887986 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.070486 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.104633 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.906223 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerStarted","Data":"dbb1e1a8cd6d5d1a9f4530cd0010ad0d48176f038fd78057908457500635e72c"} Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.907114 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-central-agent" containerID="cri-o://d2577258b0ab86ef7614d4daf96710022ca9c392c23d14e6d0e2f620ea7fe61b" gracePeriod=30 Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.907435 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.907494 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="sg-core" containerID="cri-o://5a1c2c505559b89586bede407c9b6b585d862281994a05c348540b3f34a85029" gracePeriod=30 Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.907531 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="proxy-httpd" containerID="cri-o://dbb1e1a8cd6d5d1a9f4530cd0010ad0d48176f038fd78057908457500635e72c" gracePeriod=30 Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.907518 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-notification-agent" containerID="cri-o://6d9ca2d1e938311fff282467d18b766a3e663e814778d34847a5a71b880d22c7" gracePeriod=30 Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.951655 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.950193009 podStartE2EDuration="5.951615179s" podCreationTimestamp="2026-03-09 18:49:21 +0000 UTC" firstStartedPulling="2026-03-09 18:49:22.82201113 +0000 UTC m=+1444.164483528" lastFinishedPulling="2026-03-09 18:49:25.8234333 +0000 UTC m=+1447.165905698" observedRunningTime="2026-03-09 18:49:26.942477569 +0000 UTC m=+1448.284949967" watchObservedRunningTime="2026-03-09 18:49:26.951615179 +0000 UTC m=+1448.294087577" Mar 09 18:49:26 crc kubenswrapper[4750]: I0309 18:49:26.955196 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Mar 09 18:49:27 crc kubenswrapper[4750]: I0309 18:49:27.919249 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerDied","Data":"dbb1e1a8cd6d5d1a9f4530cd0010ad0d48176f038fd78057908457500635e72c"} Mar 09 18:49:27 crc kubenswrapper[4750]: I0309 18:49:27.919198 4750 generic.go:334] "Generic (PLEG): container finished" podID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerID="dbb1e1a8cd6d5d1a9f4530cd0010ad0d48176f038fd78057908457500635e72c" exitCode=0 Mar 09 18:49:27 crc kubenswrapper[4750]: I0309 18:49:27.919616 4750 generic.go:334] "Generic (PLEG): container finished" podID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerID="5a1c2c505559b89586bede407c9b6b585d862281994a05c348540b3f34a85029" exitCode=2 Mar 09 18:49:27 crc kubenswrapper[4750]: I0309 18:49:27.919658 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerDied","Data":"5a1c2c505559b89586bede407c9b6b585d862281994a05c348540b3f34a85029"} Mar 09 18:49:27 crc kubenswrapper[4750]: I0309 18:49:27.919674 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerDied","Data":"6d9ca2d1e938311fff282467d18b766a3e663e814778d34847a5a71b880d22c7"} Mar 09 18:49:27 crc kubenswrapper[4750]: I0309 18:49:27.919696 4750 generic.go:334] "Generic (PLEG): container finished" podID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerID="6d9ca2d1e938311fff282467d18b766a3e663e814778d34847a5a71b880d22c7" exitCode=0 Mar 09 18:49:28 crc kubenswrapper[4750]: I0309 18:49:28.195332 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:28 crc kubenswrapper[4750]: I0309 18:49:28.245367 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:28 crc kubenswrapper[4750]: I0309 18:49:28.932281 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:28 crc kubenswrapper[4750]: I0309 18:49:28.966445 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Mar 09 18:49:31 crc kubenswrapper[4750]: I0309 18:49:31.976352 4750 generic.go:334] "Generic (PLEG): container finished" podID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerID="d2577258b0ab86ef7614d4daf96710022ca9c392c23d14e6d0e2f620ea7fe61b" exitCode=0 Mar 09 18:49:31 crc kubenswrapper[4750]: I0309 18:49:31.977291 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerDied","Data":"d2577258b0ab86ef7614d4daf96710022ca9c392c23d14e6d0e2f620ea7fe61b"} Mar 09 18:49:32 crc kubenswrapper[4750]: E0309 18:49:32.092008 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd293d2f1_e7ca_4d8e_b318_edbacd36fb9c.slice/crio-conmon-d2577258b0ab86ef7614d4daf96710022ca9c392c23d14e6d0e2f620ea7fe61b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd293d2f1_e7ca_4d8e_b318_edbacd36fb9c.slice/crio-d2577258b0ab86ef7614d4daf96710022ca9c392c23d14e6d0e2f620ea7fe61b.scope\": RecentStats: unable to find data in memory cache]" Mar 09 18:49:32 crc kubenswrapper[4750]: E0309 18:49:32.332077 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:32 crc kubenswrapper[4750]: E0309 18:49:32.336284 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.337458 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:32 crc kubenswrapper[4750]: E0309 18:49:32.337962 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:32 crc kubenswrapper[4750]: E0309 18:49:32.338010 4750 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.442512 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4zzp\" (UniqueName: \"kubernetes.io/projected/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-kube-api-access-f4zzp\") pod \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.442706 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-combined-ca-bundle\") pod \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.442934 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-config-data\") pod \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.442960 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-scripts\") pod \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.443049 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-sg-core-conf-yaml\") pod \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.443078 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-run-httpd\") pod \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.443107 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-log-httpd\") pod \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\" (UID: \"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c\") " Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.444013 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" (UID: "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.444089 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" (UID: "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.450685 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-kube-api-access-f4zzp" (OuterVolumeSpecName: "kube-api-access-f4zzp") pod "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" (UID: "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c"). InnerVolumeSpecName "kube-api-access-f4zzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.450884 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-scripts" (OuterVolumeSpecName: "scripts") pod "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" (UID: "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.484985 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" (UID: "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.539751 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" (UID: "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.545712 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.545955 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.545986 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.545999 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.546012 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4zzp\" (UniqueName: \"kubernetes.io/projected/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-kube-api-access-f4zzp\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.546026 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.557481 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-config-data" (OuterVolumeSpecName: "config-data") pod "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" (UID: "d293d2f1-e7ca-4d8e-b318-edbacd36fb9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.647578 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.996341 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d293d2f1-e7ca-4d8e-b318-edbacd36fb9c","Type":"ContainerDied","Data":"5f36305c574a70eee9ac5a12003b5a4bb109346141b9e6f7f853226cae1518c2"} Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.996435 4750 scope.go:117] "RemoveContainer" containerID="dbb1e1a8cd6d5d1a9f4530cd0010ad0d48176f038fd78057908457500635e72c" Mar 09 18:49:32 crc kubenswrapper[4750]: I0309 18:49:32.997860 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.051875 4750 scope.go:117] "RemoveContainer" containerID="5a1c2c505559b89586bede407c9b6b585d862281994a05c348540b3f34a85029" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.075019 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.086956 4750 scope.go:117] "RemoveContainer" containerID="6d9ca2d1e938311fff282467d18b766a3e663e814778d34847a5a71b880d22c7" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.110164 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.118347 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:33 crc kubenswrapper[4750]: E0309 18:49:33.118931 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="sg-core" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.118952 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="sg-core" Mar 09 18:49:33 crc kubenswrapper[4750]: E0309 18:49:33.118982 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="proxy-httpd" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.118991 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="proxy-httpd" Mar 09 18:49:33 crc kubenswrapper[4750]: E0309 18:49:33.119004 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-central-agent" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.119012 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-central-agent" Mar 09 18:49:33 crc kubenswrapper[4750]: E0309 18:49:33.119037 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-notification-agent" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.119043 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-notification-agent" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.119238 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-notification-agent" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.119261 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="ceilometer-central-agent" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.119275 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="sg-core" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.119293 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" containerName="proxy-httpd" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.119894 4750 scope.go:117] "RemoveContainer" containerID="d2577258b0ab86ef7614d4daf96710022ca9c392c23d14e6d0e2f620ea7fe61b" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.123069 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.125766 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.127646 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.130973 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.261517 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-scripts\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.261600 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-run-httpd\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.261678 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-config-data\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.261708 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-log-httpd\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.261777 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp6k2\" (UniqueName: \"kubernetes.io/projected/255104fc-59df-43bc-85c6-5647dc5794f5-kube-api-access-lp6k2\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.262086 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.262174 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.364889 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp6k2\" (UniqueName: \"kubernetes.io/projected/255104fc-59df-43bc-85c6-5647dc5794f5-kube-api-access-lp6k2\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.364998 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.365031 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.365131 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-scripts\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.365173 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-run-httpd\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.365225 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-config-data\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.365256 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-log-httpd\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.366009 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-log-httpd\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.366071 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-run-httpd\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.372712 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.378888 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.395485 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-config-data\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.405178 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d293d2f1-e7ca-4d8e-b318-edbacd36fb9c" path="/var/lib/kubelet/pods/d293d2f1-e7ca-4d8e-b318-edbacd36fb9c/volumes" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.407550 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp6k2\" (UniqueName: \"kubernetes.io/projected/255104fc-59df-43bc-85c6-5647dc5794f5-kube-api-access-lp6k2\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.437493 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-scripts\") pod \"ceilometer-0\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " pod="openstack/ceilometer-0" Mar 09 18:49:33 crc kubenswrapper[4750]: I0309 18:49:33.447527 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:49:34 crc kubenswrapper[4750]: I0309 18:49:34.029770 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:49:35 crc kubenswrapper[4750]: I0309 18:49:35.050097 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerStarted","Data":"e114ea9de346be1a6c965a7d381855f856a8b01559dae312b0ca63f66d40a943"} Mar 09 18:49:35 crc kubenswrapper[4750]: I0309 18:49:35.050485 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerStarted","Data":"7bdebeed7c93e60f4e165574422f8460b0569fd00e0523f2a449caa2bfea109e"} Mar 09 18:49:35 crc kubenswrapper[4750]: I0309 18:49:35.050593 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerStarted","Data":"a88cb42f417f2d61a0bb4e3f4ce3462480b199544bfd688b7a48c8a5a75cd473"} Mar 09 18:49:36 crc kubenswrapper[4750]: I0309 18:49:36.061968 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerStarted","Data":"da879d96e48d3971a4c63071806e9e803f681e393b837fd3b0920b9ea1ab3a64"} Mar 09 18:49:37 crc kubenswrapper[4750]: E0309 18:49:37.330915 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:37 crc kubenswrapper[4750]: E0309 18:49:37.333398 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:37 crc kubenswrapper[4750]: E0309 18:49:37.335381 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:37 crc kubenswrapper[4750]: E0309 18:49:37.335425 4750 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:39 crc kubenswrapper[4750]: I0309 18:49:39.093750 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerStarted","Data":"82fb87046ef59c5c5593c34fa90a82c81f7dfcb97a0ead459395e714f2a9bfe8"} Mar 09 18:49:39 crc kubenswrapper[4750]: I0309 18:49:39.094512 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:49:42 crc kubenswrapper[4750]: E0309 18:49:42.330713 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:42 crc kubenswrapper[4750]: E0309 18:49:42.332921 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:42 crc kubenswrapper[4750]: E0309 18:49:42.334466 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:42 crc kubenswrapper[4750]: E0309 18:49:42.334506 4750 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:47 crc kubenswrapper[4750]: E0309 18:49:47.337513 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:47 crc kubenswrapper[4750]: E0309 18:49:47.347791 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:47 crc kubenswrapper[4750]: E0309 18:49:47.352146 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:47 crc kubenswrapper[4750]: E0309 18:49:47.352362 4750 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.258622 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=13.310981896 podStartE2EDuration="17.258602996s" podCreationTimestamp="2026-03-09 18:49:33 +0000 UTC" firstStartedPulling="2026-03-09 18:49:34.036799816 +0000 UTC m=+1455.379272214" lastFinishedPulling="2026-03-09 18:49:37.984420916 +0000 UTC m=+1459.326893314" observedRunningTime="2026-03-09 18:49:39.12641523 +0000 UTC m=+1460.468887628" watchObservedRunningTime="2026-03-09 18:49:50.258602996 +0000 UTC m=+1471.601075394" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.266952 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fsdkk"] Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.268913 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.294807 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fsdkk"] Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.374153 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-utilities\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.374209 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tsp4\" (UniqueName: \"kubernetes.io/projected/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-kube-api-access-5tsp4\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.374234 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-catalog-content\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.476186 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-utilities\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.476295 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tsp4\" (UniqueName: \"kubernetes.io/projected/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-kube-api-access-5tsp4\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.476320 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-catalog-content\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.476854 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-utilities\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.476912 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-catalog-content\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.497283 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tsp4\" (UniqueName: \"kubernetes.io/projected/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-kube-api-access-5tsp4\") pod \"redhat-operators-fsdkk\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:50 crc kubenswrapper[4750]: I0309 18:49:50.630002 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:49:51 crc kubenswrapper[4750]: I0309 18:49:51.126744 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fsdkk"] Mar 09 18:49:51 crc kubenswrapper[4750]: I0309 18:49:51.230089 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fsdkk" event={"ID":"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2","Type":"ContainerStarted","Data":"e24f4ac88d68acc9105bed79fe01177e71c281cb3623efd53b4b1f0dde8d6bcf"} Mar 09 18:49:51 crc kubenswrapper[4750]: I0309 18:49:51.744032 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:49:51 crc kubenswrapper[4750]: I0309 18:49:51.744495 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:49:52 crc kubenswrapper[4750]: I0309 18:49:52.245080 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerID="7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e" exitCode=0 Mar 09 18:49:52 crc kubenswrapper[4750]: I0309 18:49:52.245128 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fsdkk" event={"ID":"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2","Type":"ContainerDied","Data":"7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e"} Mar 09 18:49:52 crc kubenswrapper[4750]: E0309 18:49:52.330345 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:52 crc kubenswrapper[4750]: E0309 18:49:52.331857 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:52 crc kubenswrapper[4750]: E0309 18:49:52.332975 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 09 18:49:52 crc kubenswrapper[4750]: E0309 18:49:52.333010 4750 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:53 crc kubenswrapper[4750]: I0309 18:49:53.259319 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fsdkk" event={"ID":"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2","Type":"ContainerStarted","Data":"2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653"} Mar 09 18:49:55 crc kubenswrapper[4750]: I0309 18:49:55.285663 4750 generic.go:334] "Generic (PLEG): container finished" podID="d2209408-2992-4bc3-b338-464fbaccf032" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" exitCode=137 Mar 09 18:49:55 crc kubenswrapper[4750]: I0309 18:49:55.287969 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d2209408-2992-4bc3-b338-464fbaccf032","Type":"ContainerDied","Data":"69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb"} Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.242605 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.311741 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgmw7\" (UniqueName: \"kubernetes.io/projected/d2209408-2992-4bc3-b338-464fbaccf032-kube-api-access-wgmw7\") pod \"d2209408-2992-4bc3-b338-464fbaccf032\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.311828 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-config-data\") pod \"d2209408-2992-4bc3-b338-464fbaccf032\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.311935 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-combined-ca-bundle\") pod \"d2209408-2992-4bc3-b338-464fbaccf032\" (UID: \"d2209408-2992-4bc3-b338-464fbaccf032\") " Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.321019 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2209408-2992-4bc3-b338-464fbaccf032-kube-api-access-wgmw7" (OuterVolumeSpecName: "kube-api-access-wgmw7") pod "d2209408-2992-4bc3-b338-464fbaccf032" (UID: "d2209408-2992-4bc3-b338-464fbaccf032"). InnerVolumeSpecName "kube-api-access-wgmw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.333821 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d2209408-2992-4bc3-b338-464fbaccf032","Type":"ContainerDied","Data":"50917a2fd4c630bfcf6ab88531842a9197035648e327f9e54612a136167431a6"} Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.333947 4750 scope.go:117] "RemoveContainer" containerID="69fd2b1d4d7ad16188e4ba9f26aab17986a4517bd2132c33137b731f6bd43ffb" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.334158 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.346081 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-config-data" (OuterVolumeSpecName: "config-data") pod "d2209408-2992-4bc3-b338-464fbaccf032" (UID: "d2209408-2992-4bc3-b338-464fbaccf032"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.347576 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerID="2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653" exitCode=0 Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.347654 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fsdkk" event={"ID":"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2","Type":"ContainerDied","Data":"2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653"} Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.371618 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2209408-2992-4bc3-b338-464fbaccf032" (UID: "d2209408-2992-4bc3-b338-464fbaccf032"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.415409 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgmw7\" (UniqueName: \"kubernetes.io/projected/d2209408-2992-4bc3-b338-464fbaccf032-kube-api-access-wgmw7\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.415468 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.415478 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2209408-2992-4bc3-b338-464fbaccf032-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.680978 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.698069 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.718383 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:56 crc kubenswrapper[4750]: E0309 18:49:56.719436 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.719467 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.719751 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2209408-2992-4bc3-b338-464fbaccf032" containerName="nova-cell0-conductor-conductor" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.720787 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.724701 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kljb7" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.725089 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.739384 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.827025 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0810e68-2041-4716-bcf1-97bd6c2c44aa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.827218 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwd2l\" (UniqueName: \"kubernetes.io/projected/e0810e68-2041-4716-bcf1-97bd6c2c44aa-kube-api-access-kwd2l\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.827253 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0810e68-2041-4716-bcf1-97bd6c2c44aa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.928865 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwd2l\" (UniqueName: \"kubernetes.io/projected/e0810e68-2041-4716-bcf1-97bd6c2c44aa-kube-api-access-kwd2l\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.928924 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0810e68-2041-4716-bcf1-97bd6c2c44aa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.928990 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0810e68-2041-4716-bcf1-97bd6c2c44aa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.934910 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0810e68-2041-4716-bcf1-97bd6c2c44aa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.939575 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0810e68-2041-4716-bcf1-97bd6c2c44aa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:56 crc kubenswrapper[4750]: I0309 18:49:56.948082 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwd2l\" (UniqueName: \"kubernetes.io/projected/e0810e68-2041-4716-bcf1-97bd6c2c44aa-kube-api-access-kwd2l\") pod \"nova-cell0-conductor-0\" (UID: \"e0810e68-2041-4716-bcf1-97bd6c2c44aa\") " pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:57 crc kubenswrapper[4750]: I0309 18:49:57.132237 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:57 crc kubenswrapper[4750]: I0309 18:49:57.392347 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2209408-2992-4bc3-b338-464fbaccf032" path="/var/lib/kubelet/pods/d2209408-2992-4bc3-b338-464fbaccf032/volumes" Mar 09 18:49:57 crc kubenswrapper[4750]: I0309 18:49:57.747544 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 09 18:49:57 crc kubenswrapper[4750]: W0309 18:49:57.762995 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0810e68_2041_4716_bcf1_97bd6c2c44aa.slice/crio-ce15f821e6e57da70ad8e282a406d05f18fa4b96a573f691708de78174ea08fe WatchSource:0}: Error finding container ce15f821e6e57da70ad8e282a406d05f18fa4b96a573f691708de78174ea08fe: Status 404 returned error can't find the container with id ce15f821e6e57da70ad8e282a406d05f18fa4b96a573f691708de78174ea08fe Mar 09 18:49:58 crc kubenswrapper[4750]: I0309 18:49:58.378568 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fsdkk" event={"ID":"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2","Type":"ContainerStarted","Data":"85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae"} Mar 09 18:49:58 crc kubenswrapper[4750]: I0309 18:49:58.380814 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e0810e68-2041-4716-bcf1-97bd6c2c44aa","Type":"ContainerStarted","Data":"462b069373a068eb7758aea13c9ca50e1765d224cb05db172a33d96ce3353eff"} Mar 09 18:49:58 crc kubenswrapper[4750]: I0309 18:49:58.380892 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e0810e68-2041-4716-bcf1-97bd6c2c44aa","Type":"ContainerStarted","Data":"ce15f821e6e57da70ad8e282a406d05f18fa4b96a573f691708de78174ea08fe"} Mar 09 18:49:58 crc kubenswrapper[4750]: I0309 18:49:58.380961 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 09 18:49:58 crc kubenswrapper[4750]: I0309 18:49:58.405425 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fsdkk" podStartSLOduration=3.401410005 podStartE2EDuration="8.405399219s" podCreationTimestamp="2026-03-09 18:49:50 +0000 UTC" firstStartedPulling="2026-03-09 18:49:52.248299769 +0000 UTC m=+1473.590772167" lastFinishedPulling="2026-03-09 18:49:57.252288973 +0000 UTC m=+1478.594761381" observedRunningTime="2026-03-09 18:49:58.397657599 +0000 UTC m=+1479.740129997" watchObservedRunningTime="2026-03-09 18:49:58.405399219 +0000 UTC m=+1479.747871617" Mar 09 18:49:58 crc kubenswrapper[4750]: I0309 18:49:58.430010 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.429985819 podStartE2EDuration="2.429985819s" podCreationTimestamp="2026-03-09 18:49:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:49:58.418982139 +0000 UTC m=+1479.761454547" watchObservedRunningTime="2026-03-09 18:49:58.429985819 +0000 UTC m=+1479.772458217" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.154088 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551370-bc4wr"] Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.156775 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.160068 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.160856 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.161115 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.166562 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551370-bc4wr"] Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.329959 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgfnh\" (UniqueName: \"kubernetes.io/projected/38bfb493-9d48-470c-b08f-b792cad68e98-kube-api-access-fgfnh\") pod \"auto-csr-approver-29551370-bc4wr\" (UID: \"38bfb493-9d48-470c-b08f-b792cad68e98\") " pod="openshift-infra/auto-csr-approver-29551370-bc4wr" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.431368 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgfnh\" (UniqueName: \"kubernetes.io/projected/38bfb493-9d48-470c-b08f-b792cad68e98-kube-api-access-fgfnh\") pod \"auto-csr-approver-29551370-bc4wr\" (UID: \"38bfb493-9d48-470c-b08f-b792cad68e98\") " pod="openshift-infra/auto-csr-approver-29551370-bc4wr" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.452264 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgfnh\" (UniqueName: \"kubernetes.io/projected/38bfb493-9d48-470c-b08f-b792cad68e98-kube-api-access-fgfnh\") pod \"auto-csr-approver-29551370-bc4wr\" (UID: \"38bfb493-9d48-470c-b08f-b792cad68e98\") " pod="openshift-infra/auto-csr-approver-29551370-bc4wr" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.512327 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.630597 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.630667 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:50:00 crc kubenswrapper[4750]: I0309 18:50:00.981392 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551370-bc4wr"] Mar 09 18:50:01 crc kubenswrapper[4750]: I0309 18:50:01.433826 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" event={"ID":"38bfb493-9d48-470c-b08f-b792cad68e98","Type":"ContainerStarted","Data":"bd42351ebe8e7b4f40a51ffc88a0f5ddb963c170ba5e6522702eb2258c692830"} Mar 09 18:50:01 crc kubenswrapper[4750]: I0309 18:50:01.731020 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fsdkk" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="registry-server" probeResult="failure" output=< Mar 09 18:50:01 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 18:50:01 crc kubenswrapper[4750]: > Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.165015 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.650373 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vznp2"] Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.652471 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.654698 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.654741 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.667811 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vznp2"] Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.793904 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.794278 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-scripts\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.794337 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-config-data\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.794447 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjmgk\" (UniqueName: \"kubernetes.io/projected/7df584b8-cb6c-4044-a738-8500b2a3e78b-kube-api-access-gjmgk\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.898211 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjmgk\" (UniqueName: \"kubernetes.io/projected/7df584b8-cb6c-4044-a738-8500b2a3e78b-kube-api-access-gjmgk\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.898305 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.898411 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-scripts\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.898489 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-config-data\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.914524 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-scripts\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.918012 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-config-data\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.936328 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.955990 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjmgk\" (UniqueName: \"kubernetes.io/projected/7df584b8-cb6c-4044-a738-8500b2a3e78b-kube-api-access-gjmgk\") pod \"nova-cell0-cell-mapping-vznp2\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:02 crc kubenswrapper[4750]: I0309 18:50:02.989231 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.114693 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.116488 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.124374 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.125714 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.142101 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.142305 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.213104 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.222450 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.222521 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjg9h\" (UniqueName: \"kubernetes.io/projected/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-kube-api-access-tjg9h\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.222548 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.222575 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-config-data\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.222606 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk2lv\" (UniqueName: \"kubernetes.io/projected/9e2a829f-ac14-4141-a741-9ad6c1615f52-kube-api-access-qk2lv\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.222693 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2a829f-ac14-4141-a741-9ad6c1615f52-logs\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.222959 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-config-data\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.283506 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.326841 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.326891 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjg9h\" (UniqueName: \"kubernetes.io/projected/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-kube-api-access-tjg9h\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.333616 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.333715 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.333821 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-config-data\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.333863 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk2lv\" (UniqueName: \"kubernetes.io/projected/9e2a829f-ac14-4141-a741-9ad6c1615f52-kube-api-access-qk2lv\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.334034 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2a829f-ac14-4141-a741-9ad6c1615f52-logs\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.334221 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-config-data\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.334912 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2a829f-ac14-4141-a741-9ad6c1615f52-logs\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.338557 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-config-data\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.339673 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.340069 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-config-data\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.387318 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk2lv\" (UniqueName: \"kubernetes.io/projected/9e2a829f-ac14-4141-a741-9ad6c1615f52-kube-api-access-qk2lv\") pod \"nova-api-0\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.400064 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjg9h\" (UniqueName: \"kubernetes.io/projected/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-kube-api-access-tjg9h\") pod \"nova-scheduler-0\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.401803 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.421873 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.424603 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.459922 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.478203 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.485648 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.549772 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.550406 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2srzz\" (UniqueName: \"kubernetes.io/projected/f746758c-2dda-48c8-aead-fad027da9e6e-kube-api-access-2srzz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.550538 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.565532 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.631435 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.634482 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.640468 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.669770 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2srzz\" (UniqueName: \"kubernetes.io/projected/f746758c-2dda-48c8-aead-fad027da9e6e-kube-api-access-2srzz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.669908 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.670133 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.694446 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.704125 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.705199 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.715578 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2srzz\" (UniqueName: \"kubernetes.io/projected/f746758c-2dda-48c8-aead-fad027da9e6e-kube-api-access-2srzz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.730795 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd64c67c7-xbzkw"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.744380 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.759684 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.766698 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd64c67c7-xbzkw"] Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.772952 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.773034 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-config-data\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.773053 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1018ade8-9beb-4fd4-9d10-6619b465fa96-logs\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.773078 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bft6x\" (UniqueName: \"kubernetes.io/projected/1018ade8-9beb-4fd4-9d10-6619b465fa96-kube-api-access-bft6x\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877668 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-config\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877776 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbj22\" (UniqueName: \"kubernetes.io/projected/c5fbd87e-29ab-48a6-80a2-80a905d91607-kube-api-access-mbj22\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877822 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-nb\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877874 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877898 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-svc\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-config-data\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-sb\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877957 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1018ade8-9beb-4fd4-9d10-6619b465fa96-logs\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.877983 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bft6x\" (UniqueName: \"kubernetes.io/projected/1018ade8-9beb-4fd4-9d10-6619b465fa96-kube-api-access-bft6x\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.878007 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-swift-storage-0\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.882393 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1018ade8-9beb-4fd4-9d10-6619b465fa96-logs\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.884966 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.885812 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-config-data\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.923567 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bft6x\" (UniqueName: \"kubernetes.io/projected/1018ade8-9beb-4fd4-9d10-6619b465fa96-kube-api-access-bft6x\") pod \"nova-metadata-0\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " pod="openstack/nova-metadata-0" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.986328 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbj22\" (UniqueName: \"kubernetes.io/projected/c5fbd87e-29ab-48a6-80a2-80a905d91607-kube-api-access-mbj22\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.986570 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-nb\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.986652 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-svc\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.986682 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-sb\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.986726 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-swift-storage-0\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.986819 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-config\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.987819 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-config\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.988073 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-svc\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.988825 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-nb\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:03 crc kubenswrapper[4750]: I0309 18:50:03.989464 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-swift-storage-0\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.006194 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-sb\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.008262 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.035569 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbj22\" (UniqueName: \"kubernetes.io/projected/c5fbd87e-29ab-48a6-80a2-80a905d91607-kube-api-access-mbj22\") pod \"dnsmasq-dns-cd64c67c7-xbzkw\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.109961 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.158738 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vznp2"] Mar 09 18:50:04 crc kubenswrapper[4750]: W0309 18:50:04.247322 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7df584b8_cb6c_4044_a738_8500b2a3e78b.slice/crio-0a3b1cd2f980b854d0d7fe5b62297edea78f214967f6bd0d2222ea32b10ba79c WatchSource:0}: Error finding container 0a3b1cd2f980b854d0d7fe5b62297edea78f214967f6bd0d2222ea32b10ba79c: Status 404 returned error can't find the container with id 0a3b1cd2f980b854d0d7fe5b62297edea78f214967f6bd0d2222ea32b10ba79c Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.521043 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vznp2" event={"ID":"7df584b8-cb6c-4044-a738-8500b2a3e78b","Type":"ContainerStarted","Data":"0a3b1cd2f980b854d0d7fe5b62297edea78f214967f6bd0d2222ea32b10ba79c"} Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.550002 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.708928 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l99gb"] Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.710731 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.721120 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.726012 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.747445 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l99gb"] Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.790850 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.809155 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.810478 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-scripts\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.810525 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-config-data\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.810621 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-597mq\" (UniqueName: \"kubernetes.io/projected/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-kube-api-access-597mq\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.810685 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.877435 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.921127 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-scripts\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.921173 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-config-data\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.921238 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-597mq\" (UniqueName: \"kubernetes.io/projected/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-kube-api-access-597mq\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.921269 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.932878 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-config-data\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.935235 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.944540 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-scripts\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:04 crc kubenswrapper[4750]: I0309 18:50:04.948091 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-597mq\" (UniqueName: \"kubernetes.io/projected/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-kube-api-access-597mq\") pod \"nova-cell1-conductor-db-sync-l99gb\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.020383 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd64c67c7-xbzkw"] Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.030442 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:05 crc kubenswrapper[4750]: W0309 18:50:05.086986 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5fbd87e_29ab_48a6_80a2_80a905d91607.slice/crio-74f993dbf67be49ad42c316ac514dba7768a3181b93106193a2cbdf294b11d36 WatchSource:0}: Error finding container 74f993dbf67be49ad42c316ac514dba7768a3181b93106193a2cbdf294b11d36: Status 404 returned error can't find the container with id 74f993dbf67be49ad42c316ac514dba7768a3181b93106193a2cbdf294b11d36 Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.568749 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9","Type":"ContainerStarted","Data":"badd727d8cbfbd2ed0f63bab07fff118fb3577ed90b042164dd07aefd0ad4a27"} Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.594001 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f746758c-2dda-48c8-aead-fad027da9e6e","Type":"ContainerStarted","Data":"b6cfa6e1727c2ed1ddb52d0bcbc8c61c294ea8f91164bd2ac1b35d3802777d44"} Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.621917 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1018ade8-9beb-4fd4-9d10-6619b465fa96","Type":"ContainerStarted","Data":"ee7898a86e6031ef2af0e0df1a9008190910113e93078ba3b4cbcd943879f68a"} Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.635905 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vznp2" event={"ID":"7df584b8-cb6c-4044-a738-8500b2a3e78b","Type":"ContainerStarted","Data":"6673887b27d13a728a658ddbf971f33ab76cf69b7c031681d03daffb6b374c34"} Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.664113 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" event={"ID":"38bfb493-9d48-470c-b08f-b792cad68e98","Type":"ContainerStarted","Data":"0464bab5fa574cfb5bba01a88ed142ed4f2c024f773530f5d5b22f79d500ba13"} Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.666413 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e2a829f-ac14-4141-a741-9ad6c1615f52","Type":"ContainerStarted","Data":"fc326e617fb79fcf0e6fc4c7c2dcaa6bc0182377376c34ed28b0a15285fe256e"} Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.670867 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" event={"ID":"c5fbd87e-29ab-48a6-80a2-80a905d91607","Type":"ContainerStarted","Data":"74f993dbf67be49ad42c316ac514dba7768a3181b93106193a2cbdf294b11d36"} Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.673134 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vznp2" podStartSLOduration=3.673112818 podStartE2EDuration="3.673112818s" podCreationTimestamp="2026-03-09 18:50:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:05.658666545 +0000 UTC m=+1487.001138943" watchObservedRunningTime="2026-03-09 18:50:05.673112818 +0000 UTC m=+1487.015585216" Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.688191 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" podStartSLOduration=2.873218198 podStartE2EDuration="5.688168098s" podCreationTimestamp="2026-03-09 18:50:00 +0000 UTC" firstStartedPulling="2026-03-09 18:50:00.980022564 +0000 UTC m=+1482.322494962" lastFinishedPulling="2026-03-09 18:50:03.794972464 +0000 UTC m=+1485.137444862" observedRunningTime="2026-03-09 18:50:05.684014825 +0000 UTC m=+1487.026487223" watchObservedRunningTime="2026-03-09 18:50:05.688168098 +0000 UTC m=+1487.030640506" Mar 09 18:50:05 crc kubenswrapper[4750]: I0309 18:50:05.727487 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l99gb"] Mar 09 18:50:05 crc kubenswrapper[4750]: W0309 18:50:05.749797 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fa7e3dd_d07a_48ed_8502_ef44fc7dcaa0.slice/crio-c8932a448a06807f40d3c4df7f8838c923d1f6c47dabb9116bdcb071c666c011 WatchSource:0}: Error finding container c8932a448a06807f40d3c4df7f8838c923d1f6c47dabb9116bdcb071c666c011: Status 404 returned error can't find the container with id c8932a448a06807f40d3c4df7f8838c923d1f6c47dabb9116bdcb071c666c011 Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.700855 4750 generic.go:334] "Generic (PLEG): container finished" podID="38bfb493-9d48-470c-b08f-b792cad68e98" containerID="0464bab5fa574cfb5bba01a88ed142ed4f2c024f773530f5d5b22f79d500ba13" exitCode=0 Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.701211 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" event={"ID":"38bfb493-9d48-470c-b08f-b792cad68e98","Type":"ContainerDied","Data":"0464bab5fa574cfb5bba01a88ed142ed4f2c024f773530f5d5b22f79d500ba13"} Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.718368 4750 generic.go:334] "Generic (PLEG): container finished" podID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerID="464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7" exitCode=0 Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.718467 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" event={"ID":"c5fbd87e-29ab-48a6-80a2-80a905d91607","Type":"ContainerDied","Data":"464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7"} Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.725363 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l99gb" event={"ID":"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0","Type":"ContainerStarted","Data":"f4d56c47e3e3f39a3e1ff5e639760480fdf70839f876a79fe77b5fd26dc63c3e"} Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.725407 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l99gb" event={"ID":"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0","Type":"ContainerStarted","Data":"c8932a448a06807f40d3c4df7f8838c923d1f6c47dabb9116bdcb071c666c011"} Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.774107 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-l99gb" podStartSLOduration=2.774085275 podStartE2EDuration="2.774085275s" podCreationTimestamp="2026-03-09 18:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:06.768883944 +0000 UTC m=+1488.111356352" watchObservedRunningTime="2026-03-09 18:50:06.774085275 +0000 UTC m=+1488.116557673" Mar 09 18:50:06 crc kubenswrapper[4750]: I0309 18:50:06.980706 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:07 crc kubenswrapper[4750]: I0309 18:50:07.000360 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.311248 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.312115 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="64ff0811-8670-4605-9ef6-383f43259ace" containerName="kube-state-metrics" containerID="cri-o://b0968c2bd4cee8b2f094ec3aa22750800866cfe3a01f5dd80c9ea696958866c4" gracePeriod=30 Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.454904 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.601303 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgfnh\" (UniqueName: \"kubernetes.io/projected/38bfb493-9d48-470c-b08f-b792cad68e98-kube-api-access-fgfnh\") pod \"38bfb493-9d48-470c-b08f-b792cad68e98\" (UID: \"38bfb493-9d48-470c-b08f-b792cad68e98\") " Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.622947 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38bfb493-9d48-470c-b08f-b792cad68e98-kube-api-access-fgfnh" (OuterVolumeSpecName: "kube-api-access-fgfnh") pod "38bfb493-9d48-470c-b08f-b792cad68e98" (UID: "38bfb493-9d48-470c-b08f-b792cad68e98"). InnerVolumeSpecName "kube-api-access-fgfnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.707123 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgfnh\" (UniqueName: \"kubernetes.io/projected/38bfb493-9d48-470c-b08f-b792cad68e98-kube-api-access-fgfnh\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.796805 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" event={"ID":"c5fbd87e-29ab-48a6-80a2-80a905d91607","Type":"ContainerStarted","Data":"c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e"} Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.798040 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.823008 4750 generic.go:334] "Generic (PLEG): container finished" podID="64ff0811-8670-4605-9ef6-383f43259ace" containerID="b0968c2bd4cee8b2f094ec3aa22750800866cfe3a01f5dd80c9ea696958866c4" exitCode=2 Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.823079 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"64ff0811-8670-4605-9ef6-383f43259ace","Type":"ContainerDied","Data":"b0968c2bd4cee8b2f094ec3aa22750800866cfe3a01f5dd80c9ea696958866c4"} Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.857759 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" event={"ID":"38bfb493-9d48-470c-b08f-b792cad68e98","Type":"ContainerDied","Data":"bd42351ebe8e7b4f40a51ffc88a0f5ddb963c170ba5e6522702eb2258c692830"} Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.857825 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd42351ebe8e7b4f40a51ffc88a0f5ddb963c170ba5e6522702eb2258c692830" Mar 09 18:50:10 crc kubenswrapper[4750]: I0309 18:50:10.858062 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551370-bc4wr" Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.372071 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.406880 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" podStartSLOduration=8.406856626 podStartE2EDuration="8.406856626s" podCreationTimestamp="2026-03-09 18:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:10.870379414 +0000 UTC m=+1492.212851812" watchObservedRunningTime="2026-03-09 18:50:11.406856626 +0000 UTC m=+1492.749329024" Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.525911 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6vzz\" (UniqueName: \"kubernetes.io/projected/64ff0811-8670-4605-9ef6-383f43259ace-kube-api-access-z6vzz\") pod \"64ff0811-8670-4605-9ef6-383f43259ace\" (UID: \"64ff0811-8670-4605-9ef6-383f43259ace\") " Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.537926 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ff0811-8670-4605-9ef6-383f43259ace-kube-api-access-z6vzz" (OuterVolumeSpecName: "kube-api-access-z6vzz") pod "64ff0811-8670-4605-9ef6-383f43259ace" (UID: "64ff0811-8670-4605-9ef6-383f43259ace"). InnerVolumeSpecName "kube-api-access-z6vzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.562882 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551364-gv2jc"] Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.587198 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551364-gv2jc"] Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.629578 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6vzz\" (UniqueName: \"kubernetes.io/projected/64ff0811-8670-4605-9ef6-383f43259ace-kube-api-access-z6vzz\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.742682 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fsdkk" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="registry-server" probeResult="failure" output=< Mar 09 18:50:11 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 18:50:11 crc kubenswrapper[4750]: > Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.874749 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f746758c-2dda-48c8-aead-fad027da9e6e","Type":"ContainerStarted","Data":"7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4"} Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.874967 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="f746758c-2dda-48c8-aead-fad027da9e6e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4" gracePeriod=30 Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.877985 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1018ade8-9beb-4fd4-9d10-6619b465fa96","Type":"ContainerStarted","Data":"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86"} Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.878015 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1018ade8-9beb-4fd4-9d10-6619b465fa96","Type":"ContainerStarted","Data":"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1"} Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.878113 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-log" containerID="cri-o://4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1" gracePeriod=30 Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.878149 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-metadata" containerID="cri-o://9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86" gracePeriod=30 Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.881561 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.881582 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"64ff0811-8670-4605-9ef6-383f43259ace","Type":"ContainerDied","Data":"9eb1fbbffa42af84ada906b4a3bb797f2f2794dfe009274acd26470d2aec7191"} Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.881677 4750 scope.go:117] "RemoveContainer" containerID="b0968c2bd4cee8b2f094ec3aa22750800866cfe3a01f5dd80c9ea696958866c4" Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.900965 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e2a829f-ac14-4141-a741-9ad6c1615f52","Type":"ContainerStarted","Data":"5047209590a7b4f61642164ff4a9927483d2a43c6caf68e8a2aa40527b28b88e"} Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.901020 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e2a829f-ac14-4141-a741-9ad6c1615f52","Type":"ContainerStarted","Data":"1e7723029dd26e346b160c3e043bdae867f720f655aaa6001ee91e73b445d04b"} Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.918928 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9","Type":"ContainerStarted","Data":"48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474"} Mar 09 18:50:11 crc kubenswrapper[4750]: I0309 18:50:11.980502 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.6022573749999998 podStartE2EDuration="8.980380578s" podCreationTimestamp="2026-03-09 18:50:03 +0000 UTC" firstStartedPulling="2026-03-09 18:50:04.76632614 +0000 UTC m=+1486.108798538" lastFinishedPulling="2026-03-09 18:50:10.144449353 +0000 UTC m=+1491.486921741" observedRunningTime="2026-03-09 18:50:11.899953517 +0000 UTC m=+1493.242425915" watchObservedRunningTime="2026-03-09 18:50:11.980380578 +0000 UTC m=+1493.322852976" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.039647 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.782187796 podStartE2EDuration="9.03959754s" podCreationTimestamp="2026-03-09 18:50:03 +0000 UTC" firstStartedPulling="2026-03-09 18:50:04.887792289 +0000 UTC m=+1486.230264687" lastFinishedPulling="2026-03-09 18:50:10.145202033 +0000 UTC m=+1491.487674431" observedRunningTime="2026-03-09 18:50:11.923220461 +0000 UTC m=+1493.265692849" watchObservedRunningTime="2026-03-09 18:50:12.03959754 +0000 UTC m=+1493.382069938" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.075721 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.096322 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.111187 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:50:12 crc kubenswrapper[4750]: E0309 18:50:12.111917 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ff0811-8670-4605-9ef6-383f43259ace" containerName="kube-state-metrics" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.111951 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ff0811-8670-4605-9ef6-383f43259ace" containerName="kube-state-metrics" Mar 09 18:50:12 crc kubenswrapper[4750]: E0309 18:50:12.111966 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38bfb493-9d48-470c-b08f-b792cad68e98" containerName="oc" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.111976 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="38bfb493-9d48-470c-b08f-b792cad68e98" containerName="oc" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.112225 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="38bfb493-9d48-470c-b08f-b792cad68e98" containerName="oc" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.112257 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="64ff0811-8670-4605-9ef6-383f43259ace" containerName="kube-state-metrics" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.113085 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.117357 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.117350 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.738871236 podStartE2EDuration="9.117326968s" podCreationTimestamp="2026-03-09 18:50:03 +0000 UTC" firstStartedPulling="2026-03-09 18:50:04.766093144 +0000 UTC m=+1486.108565542" lastFinishedPulling="2026-03-09 18:50:10.144548876 +0000 UTC m=+1491.487021274" observedRunningTime="2026-03-09 18:50:11.983095722 +0000 UTC m=+1493.325568120" watchObservedRunningTime="2026-03-09 18:50:12.117326968 +0000 UTC m=+1493.459799366" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.117621 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.148713 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.185157 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.6387548990000003 podStartE2EDuration="9.185128674s" podCreationTimestamp="2026-03-09 18:50:03 +0000 UTC" firstStartedPulling="2026-03-09 18:50:04.5963309 +0000 UTC m=+1485.938803298" lastFinishedPulling="2026-03-09 18:50:10.142704675 +0000 UTC m=+1491.485177073" observedRunningTime="2026-03-09 18:50:12.043879207 +0000 UTC m=+1493.386351605" watchObservedRunningTime="2026-03-09 18:50:12.185128674 +0000 UTC m=+1493.527601072" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.262023 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvtfx\" (UniqueName: \"kubernetes.io/projected/47b019da-8326-4a33-80f4-257dd5517e92-kube-api-access-zvtfx\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.262103 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.262368 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:12 crc kubenswrapper[4750]: I0309 18:50:12.262447 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.364818 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.364901 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.364943 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.365141 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvtfx\" (UniqueName: \"kubernetes.io/projected/47b019da-8326-4a33-80f4-257dd5517e92-kube-api-access-zvtfx\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.371285 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.374451 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.384419 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b019da-8326-4a33-80f4-257dd5517e92-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.385299 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvtfx\" (UniqueName: \"kubernetes.io/projected/47b019da-8326-4a33-80f4-257dd5517e92-kube-api-access-zvtfx\") pod \"kube-state-metrics-0\" (UID: \"47b019da-8326-4a33-80f4-257dd5517e92\") " pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.494075 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.693196 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.773746 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-combined-ca-bundle\") pod \"1018ade8-9beb-4fd4-9d10-6619b465fa96\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.773808 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-config-data\") pod \"1018ade8-9beb-4fd4-9d10-6619b465fa96\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.773860 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1018ade8-9beb-4fd4-9d10-6619b465fa96-logs\") pod \"1018ade8-9beb-4fd4-9d10-6619b465fa96\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.773914 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bft6x\" (UniqueName: \"kubernetes.io/projected/1018ade8-9beb-4fd4-9d10-6619b465fa96-kube-api-access-bft6x\") pod \"1018ade8-9beb-4fd4-9d10-6619b465fa96\" (UID: \"1018ade8-9beb-4fd4-9d10-6619b465fa96\") " Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.776370 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1018ade8-9beb-4fd4-9d10-6619b465fa96-logs" (OuterVolumeSpecName: "logs") pod "1018ade8-9beb-4fd4-9d10-6619b465fa96" (UID: "1018ade8-9beb-4fd4-9d10-6619b465fa96"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.786884 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1018ade8-9beb-4fd4-9d10-6619b465fa96-kube-api-access-bft6x" (OuterVolumeSpecName: "kube-api-access-bft6x") pod "1018ade8-9beb-4fd4-9d10-6619b465fa96" (UID: "1018ade8-9beb-4fd4-9d10-6619b465fa96"). InnerVolumeSpecName "kube-api-access-bft6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.805820 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-config-data" (OuterVolumeSpecName: "config-data") pod "1018ade8-9beb-4fd4-9d10-6619b465fa96" (UID: "1018ade8-9beb-4fd4-9d10-6619b465fa96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.817869 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1018ade8-9beb-4fd4-9d10-6619b465fa96" (UID: "1018ade8-9beb-4fd4-9d10-6619b465fa96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.876399 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.876428 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1018ade8-9beb-4fd4-9d10-6619b465fa96-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.876438 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1018ade8-9beb-4fd4-9d10-6619b465fa96-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.876446 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bft6x\" (UniqueName: \"kubernetes.io/projected/1018ade8-9beb-4fd4-9d10-6619b465fa96-kube-api-access-bft6x\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.933481 4750 generic.go:334] "Generic (PLEG): container finished" podID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerID="9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86" exitCode=0 Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.933518 4750 generic.go:334] "Generic (PLEG): container finished" podID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerID="4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1" exitCode=143 Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.933572 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1018ade8-9beb-4fd4-9d10-6619b465fa96","Type":"ContainerDied","Data":"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86"} Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.933606 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1018ade8-9beb-4fd4-9d10-6619b465fa96","Type":"ContainerDied","Data":"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1"} Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.933619 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1018ade8-9beb-4fd4-9d10-6619b465fa96","Type":"ContainerDied","Data":"ee7898a86e6031ef2af0e0df1a9008190910113e93078ba3b4cbcd943879f68a"} Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.933661 4750 scope.go:117] "RemoveContainer" containerID="9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.933842 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.982217 4750 scope.go:117] "RemoveContainer" containerID="4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.986321 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:12.998716 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.034128 4750 scope.go:117] "RemoveContainer" containerID="9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86" Mar 09 18:50:13 crc kubenswrapper[4750]: E0309 18:50:13.034592 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86\": container with ID starting with 9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86 not found: ID does not exist" containerID="9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.034614 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86"} err="failed to get container status \"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86\": rpc error: code = NotFound desc = could not find container \"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86\": container with ID starting with 9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86 not found: ID does not exist" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.034664 4750 scope.go:117] "RemoveContainer" containerID="4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1" Mar 09 18:50:13 crc kubenswrapper[4750]: E0309 18:50:13.034883 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1\": container with ID starting with 4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1 not found: ID does not exist" containerID="4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.034903 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1"} err="failed to get container status \"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1\": rpc error: code = NotFound desc = could not find container \"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1\": container with ID starting with 4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1 not found: ID does not exist" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.034917 4750 scope.go:117] "RemoveContainer" containerID="9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.035139 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86"} err="failed to get container status \"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86\": rpc error: code = NotFound desc = could not find container \"9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86\": container with ID starting with 9ce249868026501c7afe5bedc6856ccf9866c9951902e4d12398eff40b1b5e86 not found: ID does not exist" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.035153 4750 scope.go:117] "RemoveContainer" containerID="4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.035365 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1"} err="failed to get container status \"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1\": rpc error: code = NotFound desc = could not find container \"4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1\": container with ID starting with 4bf6f3d393978e6e989c2afa470b7fb8db26ea796db2e4100d7b0eb7dfc3cba1 not found: ID does not exist" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.040721 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:13 crc kubenswrapper[4750]: E0309 18:50:13.041474 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-metadata" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.041495 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-metadata" Mar 09 18:50:13 crc kubenswrapper[4750]: E0309 18:50:13.041506 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-log" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.041513 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-log" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.041783 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-log" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.041818 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" containerName="nova-metadata-metadata" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.043202 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.046463 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.046743 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.051923 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.182368 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498246e2-9cdd-4a11-984b-3b703f2108bc-logs\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.182416 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.182460 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.182659 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-config-data\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.182679 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f55nk\" (UniqueName: \"kubernetes.io/projected/498246e2-9cdd-4a11-984b-3b703f2108bc-kube-api-access-f55nk\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.284085 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-config-data\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.284152 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f55nk\" (UniqueName: \"kubernetes.io/projected/498246e2-9cdd-4a11-984b-3b703f2108bc-kube-api-access-f55nk\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.284241 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498246e2-9cdd-4a11-984b-3b703f2108bc-logs\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.284284 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.284317 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.287308 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498246e2-9cdd-4a11-984b-3b703f2108bc-logs\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.289988 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-config-data\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.315262 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.320319 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f55nk\" (UniqueName: \"kubernetes.io/projected/498246e2-9cdd-4a11-984b-3b703f2108bc-kube-api-access-f55nk\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.323820 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.373149 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.414858 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1018ade8-9beb-4fd4-9d10-6619b465fa96" path="/var/lib/kubelet/pods/1018ade8-9beb-4fd4-9d10-6619b465fa96/volumes" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.429562 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39b5ffd7-804d-47a4-bc5b-f000f7a425a5" path="/var/lib/kubelet/pods/39b5ffd7-804d-47a4-bc5b-f000f7a425a5/volumes" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.433385 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64ff0811-8670-4605-9ef6-383f43259ace" path="/var/lib/kubelet/pods/64ff0811-8670-4605-9ef6-383f43259ace/volumes" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.461719 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.461762 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.481367 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.481732 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.523839 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 09 18:50:13 crc kubenswrapper[4750]: W0309 18:50:13.648808 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47b019da_8326_4a33_80f4_257dd5517e92.slice/crio-f904c217bf02d838492bdbc3e73cccf4cd2d7864a8e2050c635a6296dbda5ab1 WatchSource:0}: Error finding container f904c217bf02d838492bdbc3e73cccf4cd2d7864a8e2050c635a6296dbda5ab1: Status 404 returned error can't find the container with id f904c217bf02d838492bdbc3e73cccf4cd2d7864a8e2050c635a6296dbda5ab1 Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.669431 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.765775 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:13 crc kubenswrapper[4750]: I0309 18:50:13.974946 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"47b019da-8326-4a33-80f4-257dd5517e92","Type":"ContainerStarted","Data":"f904c217bf02d838492bdbc3e73cccf4cd2d7864a8e2050c635a6296dbda5ab1"} Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.057718 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.122918 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.545826 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.221:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.545826 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.221:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.921294 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.921949 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-central-agent" containerID="cri-o://7bdebeed7c93e60f4e165574422f8460b0569fd00e0523f2a449caa2bfea109e" gracePeriod=30 Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.922605 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="proxy-httpd" containerID="cri-o://82fb87046ef59c5c5593c34fa90a82c81f7dfcb97a0ead459395e714f2a9bfe8" gracePeriod=30 Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.922695 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-notification-agent" containerID="cri-o://e114ea9de346be1a6c965a7d381855f856a8b01559dae312b0ca63f66d40a943" gracePeriod=30 Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.922710 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="sg-core" containerID="cri-o://da879d96e48d3971a4c63071806e9e803f681e393b837fd3b0920b9ea1ab3a64" gracePeriod=30 Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.993396 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"47b019da-8326-4a33-80f4-257dd5517e92","Type":"ContainerStarted","Data":"9f801be581295057626fa1ca0f7cadaf5a32960f25b43b6e8e4ed0dbe3742836"} Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.993844 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.995598 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"498246e2-9cdd-4a11-984b-3b703f2108bc","Type":"ContainerStarted","Data":"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5"} Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.995648 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"498246e2-9cdd-4a11-984b-3b703f2108bc","Type":"ContainerStarted","Data":"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e"} Mar 09 18:50:14 crc kubenswrapper[4750]: I0309 18:50:14.995661 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"498246e2-9cdd-4a11-984b-3b703f2108bc","Type":"ContainerStarted","Data":"8f4d6a490026e437d49729ff8b369d2890cac5ee22ae5d33872436c78be988e3"} Mar 09 18:50:15 crc kubenswrapper[4750]: I0309 18:50:15.023647 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.496407856 podStartE2EDuration="4.023612065s" podCreationTimestamp="2026-03-09 18:50:11 +0000 UTC" firstStartedPulling="2026-03-09 18:50:13.681220503 +0000 UTC m=+1495.023692891" lastFinishedPulling="2026-03-09 18:50:14.208424692 +0000 UTC m=+1495.550897100" observedRunningTime="2026-03-09 18:50:15.014006473 +0000 UTC m=+1496.356478871" watchObservedRunningTime="2026-03-09 18:50:15.023612065 +0000 UTC m=+1496.366084463" Mar 09 18:50:15 crc kubenswrapper[4750]: I0309 18:50:15.042859 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.042835629 podStartE2EDuration="3.042835629s" podCreationTimestamp="2026-03-09 18:50:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:15.036584619 +0000 UTC m=+1496.379057017" watchObservedRunningTime="2026-03-09 18:50:15.042835629 +0000 UTC m=+1496.385308027" Mar 09 18:50:16 crc kubenswrapper[4750]: I0309 18:50:16.008228 4750 generic.go:334] "Generic (PLEG): container finished" podID="255104fc-59df-43bc-85c6-5647dc5794f5" containerID="82fb87046ef59c5c5593c34fa90a82c81f7dfcb97a0ead459395e714f2a9bfe8" exitCode=0 Mar 09 18:50:16 crc kubenswrapper[4750]: I0309 18:50:16.008564 4750 generic.go:334] "Generic (PLEG): container finished" podID="255104fc-59df-43bc-85c6-5647dc5794f5" containerID="da879d96e48d3971a4c63071806e9e803f681e393b837fd3b0920b9ea1ab3a64" exitCode=2 Mar 09 18:50:16 crc kubenswrapper[4750]: I0309 18:50:16.008579 4750 generic.go:334] "Generic (PLEG): container finished" podID="255104fc-59df-43bc-85c6-5647dc5794f5" containerID="7bdebeed7c93e60f4e165574422f8460b0569fd00e0523f2a449caa2bfea109e" exitCode=0 Mar 09 18:50:16 crc kubenswrapper[4750]: I0309 18:50:16.009732 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerDied","Data":"82fb87046ef59c5c5593c34fa90a82c81f7dfcb97a0ead459395e714f2a9bfe8"} Mar 09 18:50:16 crc kubenswrapper[4750]: I0309 18:50:16.009769 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerDied","Data":"da879d96e48d3971a4c63071806e9e803f681e393b837fd3b0920b9ea1ab3a64"} Mar 09 18:50:16 crc kubenswrapper[4750]: I0309 18:50:16.009806 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerDied","Data":"7bdebeed7c93e60f4e165574422f8460b0569fd00e0523f2a449caa2bfea109e"} Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.046604 4750 generic.go:334] "Generic (PLEG): container finished" podID="7df584b8-cb6c-4044-a738-8500b2a3e78b" containerID="6673887b27d13a728a658ddbf971f33ab76cf69b7c031681d03daffb6b374c34" exitCode=0 Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.048577 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vznp2" event={"ID":"7df584b8-cb6c-4044-a738-8500b2a3e78b","Type":"ContainerDied","Data":"6673887b27d13a728a658ddbf971f33ab76cf69b7c031681d03daffb6b374c34"} Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.062528 4750 generic.go:334] "Generic (PLEG): container finished" podID="255104fc-59df-43bc-85c6-5647dc5794f5" containerID="e114ea9de346be1a6c965a7d381855f856a8b01559dae312b0ca63f66d40a943" exitCode=0 Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.062590 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerDied","Data":"e114ea9de346be1a6c965a7d381855f856a8b01559dae312b0ca63f66d40a943"} Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.231497 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.337394 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-combined-ca-bundle\") pod \"255104fc-59df-43bc-85c6-5647dc5794f5\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.337574 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-run-httpd\") pod \"255104fc-59df-43bc-85c6-5647dc5794f5\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.337606 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp6k2\" (UniqueName: \"kubernetes.io/projected/255104fc-59df-43bc-85c6-5647dc5794f5-kube-api-access-lp6k2\") pod \"255104fc-59df-43bc-85c6-5647dc5794f5\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.337665 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-sg-core-conf-yaml\") pod \"255104fc-59df-43bc-85c6-5647dc5794f5\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.337696 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-scripts\") pod \"255104fc-59df-43bc-85c6-5647dc5794f5\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.337729 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-config-data\") pod \"255104fc-59df-43bc-85c6-5647dc5794f5\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.337884 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-log-httpd\") pod \"255104fc-59df-43bc-85c6-5647dc5794f5\" (UID: \"255104fc-59df-43bc-85c6-5647dc5794f5\") " Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.339355 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "255104fc-59df-43bc-85c6-5647dc5794f5" (UID: "255104fc-59df-43bc-85c6-5647dc5794f5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.339605 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "255104fc-59df-43bc-85c6-5647dc5794f5" (UID: "255104fc-59df-43bc-85c6-5647dc5794f5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.370937 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-scripts" (OuterVolumeSpecName: "scripts") pod "255104fc-59df-43bc-85c6-5647dc5794f5" (UID: "255104fc-59df-43bc-85c6-5647dc5794f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.370984 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/255104fc-59df-43bc-85c6-5647dc5794f5-kube-api-access-lp6k2" (OuterVolumeSpecName: "kube-api-access-lp6k2") pod "255104fc-59df-43bc-85c6-5647dc5794f5" (UID: "255104fc-59df-43bc-85c6-5647dc5794f5"). InnerVolumeSpecName "kube-api-access-lp6k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.374492 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.374603 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.379803 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "255104fc-59df-43bc-85c6-5647dc5794f5" (UID: "255104fc-59df-43bc-85c6-5647dc5794f5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.446381 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.446414 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp6k2\" (UniqueName: \"kubernetes.io/projected/255104fc-59df-43bc-85c6-5647dc5794f5-kube-api-access-lp6k2\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.446425 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.446435 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.446444 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/255104fc-59df-43bc-85c6-5647dc5794f5-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.450872 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "255104fc-59df-43bc-85c6-5647dc5794f5" (UID: "255104fc-59df-43bc-85c6-5647dc5794f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.475111 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-config-data" (OuterVolumeSpecName: "config-data") pod "255104fc-59df-43bc-85c6-5647dc5794f5" (UID: "255104fc-59df-43bc-85c6-5647dc5794f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.548649 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:18 crc kubenswrapper[4750]: I0309 18:50:18.548693 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255104fc-59df-43bc-85c6-5647dc5794f5-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.079165 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"255104fc-59df-43bc-85c6-5647dc5794f5","Type":"ContainerDied","Data":"a88cb42f417f2d61a0bb4e3f4ce3462480b199544bfd688b7a48c8a5a75cd473"} Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.079246 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.081579 4750 scope.go:117] "RemoveContainer" containerID="82fb87046ef59c5c5593c34fa90a82c81f7dfcb97a0ead459395e714f2a9bfe8" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.111958 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.115281 4750 scope.go:117] "RemoveContainer" containerID="da879d96e48d3971a4c63071806e9e803f681e393b837fd3b0920b9ea1ab3a64" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.142589 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.167817 4750 scope.go:117] "RemoveContainer" containerID="e114ea9de346be1a6c965a7d381855f856a8b01559dae312b0ca63f66d40a943" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.172515 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.192446 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:19 crc kubenswrapper[4750]: E0309 18:50:19.196978 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="proxy-httpd" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197002 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="proxy-httpd" Mar 09 18:50:19 crc kubenswrapper[4750]: E0309 18:50:19.197020 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="sg-core" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197026 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="sg-core" Mar 09 18:50:19 crc kubenswrapper[4750]: E0309 18:50:19.197040 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-central-agent" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197046 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-central-agent" Mar 09 18:50:19 crc kubenswrapper[4750]: E0309 18:50:19.197055 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-notification-agent" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197061 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-notification-agent" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197253 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-central-agent" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197268 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="ceilometer-notification-agent" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197278 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="proxy-httpd" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.197292 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" containerName="sg-core" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.201225 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.208715 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dfd9d95f-cvfsd"] Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.208985 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" podUID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerName="dnsmasq-dns" containerID="cri-o://8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784" gracePeriod=10 Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.219449 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.219747 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.219897 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.250380 4750 scope.go:117] "RemoveContainer" containerID="7bdebeed7c93e60f4e165574422f8460b0569fd00e0523f2a449caa2bfea109e" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267435 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-log-httpd\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267513 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qx5x\" (UniqueName: \"kubernetes.io/projected/221bc157-a4af-4e07-89fa-1bbd1dcd254c-kube-api-access-4qx5x\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267599 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267648 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267768 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-scripts\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267792 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-config-data\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267811 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.267831 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-run-httpd\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.270599 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370199 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-scripts\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370238 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-config-data\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370265 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370289 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-run-httpd\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370318 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-log-httpd\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370359 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qx5x\" (UniqueName: \"kubernetes.io/projected/221bc157-a4af-4e07-89fa-1bbd1dcd254c-kube-api-access-4qx5x\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370399 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.370417 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.371644 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-log-httpd\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.371887 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-run-httpd\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.375797 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.375966 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.377458 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-scripts\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.378877 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-config-data\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.388509 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.400160 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qx5x\" (UniqueName: \"kubernetes.io/projected/221bc157-a4af-4e07-89fa-1bbd1dcd254c-kube-api-access-4qx5x\") pod \"ceilometer-0\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.524339 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="255104fc-59df-43bc-85c6-5647dc5794f5" path="/var/lib/kubelet/pods/255104fc-59df-43bc-85c6-5647dc5794f5/volumes" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.549320 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.688223 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.816989 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjmgk\" (UniqueName: \"kubernetes.io/projected/7df584b8-cb6c-4044-a738-8500b2a3e78b-kube-api-access-gjmgk\") pod \"7df584b8-cb6c-4044-a738-8500b2a3e78b\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.817084 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-config-data\") pod \"7df584b8-cb6c-4044-a738-8500b2a3e78b\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.817120 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-combined-ca-bundle\") pod \"7df584b8-cb6c-4044-a738-8500b2a3e78b\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.817197 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-scripts\") pod \"7df584b8-cb6c-4044-a738-8500b2a3e78b\" (UID: \"7df584b8-cb6c-4044-a738-8500b2a3e78b\") " Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.825079 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-scripts" (OuterVolumeSpecName: "scripts") pod "7df584b8-cb6c-4044-a738-8500b2a3e78b" (UID: "7df584b8-cb6c-4044-a738-8500b2a3e78b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.824288 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df584b8-cb6c-4044-a738-8500b2a3e78b-kube-api-access-gjmgk" (OuterVolumeSpecName: "kube-api-access-gjmgk") pod "7df584b8-cb6c-4044-a738-8500b2a3e78b" (UID: "7df584b8-cb6c-4044-a738-8500b2a3e78b"). InnerVolumeSpecName "kube-api-access-gjmgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.857842 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7df584b8-cb6c-4044-a738-8500b2a3e78b" (UID: "7df584b8-cb6c-4044-a738-8500b2a3e78b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.871074 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-config-data" (OuterVolumeSpecName: "config-data") pod "7df584b8-cb6c-4044-a738-8500b2a3e78b" (UID: "7df584b8-cb6c-4044-a738-8500b2a3e78b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.922380 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.922418 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.922428 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjmgk\" (UniqueName: \"kubernetes.io/projected/7df584b8-cb6c-4044-a738-8500b2a3e78b-kube-api-access-gjmgk\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.922440 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df584b8-cb6c-4044-a738-8500b2a3e78b-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:19 crc kubenswrapper[4750]: I0309 18:50:19.992712 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.093392 4750 generic.go:334] "Generic (PLEG): container finished" podID="6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" containerID="f4d56c47e3e3f39a3e1ff5e639760480fdf70839f876a79fe77b5fd26dc63c3e" exitCode=0 Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.093477 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l99gb" event={"ID":"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0","Type":"ContainerDied","Data":"f4d56c47e3e3f39a3e1ff5e639760480fdf70839f876a79fe77b5fd26dc63c3e"} Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.104975 4750 generic.go:334] "Generic (PLEG): container finished" podID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerID="8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784" exitCode=0 Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.105685 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" event={"ID":"74f3d684-e75b-4b86-9086-e2649c75ee9e","Type":"ContainerDied","Data":"8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784"} Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.105752 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" event={"ID":"74f3d684-e75b-4b86-9086-e2649c75ee9e","Type":"ContainerDied","Data":"ad88955e1b64b42e80bf03f2bcaf1222d483a93bc20a3540b662f1dc8613d558"} Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.105772 4750 scope.go:117] "RemoveContainer" containerID="8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.107265 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfd9d95f-cvfsd" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.118698 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vznp2" event={"ID":"7df584b8-cb6c-4044-a738-8500b2a3e78b","Type":"ContainerDied","Data":"0a3b1cd2f980b854d0d7fe5b62297edea78f214967f6bd0d2222ea32b10ba79c"} Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.118739 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a3b1cd2f980b854d0d7fe5b62297edea78f214967f6bd0d2222ea32b10ba79c" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.118820 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vznp2" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.127944 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-swift-storage-0\") pod \"74f3d684-e75b-4b86-9086-e2649c75ee9e\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.127995 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-svc\") pod \"74f3d684-e75b-4b86-9086-e2649c75ee9e\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.128016 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-sb\") pod \"74f3d684-e75b-4b86-9086-e2649c75ee9e\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.128128 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-nb\") pod \"74f3d684-e75b-4b86-9086-e2649c75ee9e\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.128194 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ljvz\" (UniqueName: \"kubernetes.io/projected/74f3d684-e75b-4b86-9086-e2649c75ee9e-kube-api-access-9ljvz\") pod \"74f3d684-e75b-4b86-9086-e2649c75ee9e\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.128307 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-config\") pod \"74f3d684-e75b-4b86-9086-e2649c75ee9e\" (UID: \"74f3d684-e75b-4b86-9086-e2649c75ee9e\") " Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.142621 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74f3d684-e75b-4b86-9086-e2649c75ee9e-kube-api-access-9ljvz" (OuterVolumeSpecName: "kube-api-access-9ljvz") pod "74f3d684-e75b-4b86-9086-e2649c75ee9e" (UID: "74f3d684-e75b-4b86-9086-e2649c75ee9e"). InnerVolumeSpecName "kube-api-access-9ljvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.160855 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.164728 4750 scope.go:117] "RemoveContainer" containerID="559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.192712 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "74f3d684-e75b-4b86-9086-e2649c75ee9e" (UID: "74f3d684-e75b-4b86-9086-e2649c75ee9e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.208485 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "74f3d684-e75b-4b86-9086-e2649c75ee9e" (UID: "74f3d684-e75b-4b86-9086-e2649c75ee9e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.214778 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-config" (OuterVolumeSpecName: "config") pod "74f3d684-e75b-4b86-9086-e2649c75ee9e" (UID: "74f3d684-e75b-4b86-9086-e2649c75ee9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.216431 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "74f3d684-e75b-4b86-9086-e2649c75ee9e" (UID: "74f3d684-e75b-4b86-9086-e2649c75ee9e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.231186 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.231227 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ljvz\" (UniqueName: \"kubernetes.io/projected/74f3d684-e75b-4b86-9086-e2649c75ee9e-kube-api-access-9ljvz\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.231244 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.231257 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.231269 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.231488 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "74f3d684-e75b-4b86-9086-e2649c75ee9e" (UID: "74f3d684-e75b-4b86-9086-e2649c75ee9e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.248069 4750 scope.go:117] "RemoveContainer" containerID="8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784" Mar 09 18:50:20 crc kubenswrapper[4750]: E0309 18:50:20.248488 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784\": container with ID starting with 8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784 not found: ID does not exist" containerID="8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.248548 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784"} err="failed to get container status \"8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784\": rpc error: code = NotFound desc = could not find container \"8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784\": container with ID starting with 8f2a18d943219c785d8449adc2e85dac5c64aa0a2bf72c5871ff979b164e4784 not found: ID does not exist" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.248579 4750 scope.go:117] "RemoveContainer" containerID="559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c" Mar 09 18:50:20 crc kubenswrapper[4750]: E0309 18:50:20.248968 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c\": container with ID starting with 559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c not found: ID does not exist" containerID="559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.249008 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c"} err="failed to get container status \"559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c\": rpc error: code = NotFound desc = could not find container \"559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c\": container with ID starting with 559b397c56cda76341819ed3096a4055dcf961fbe06d7da1203ba5caa770de8c not found: ID does not exist" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.269144 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.269380 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-log" containerID="cri-o://1e7723029dd26e346b160c3e043bdae867f720f655aaa6001ee91e73b445d04b" gracePeriod=30 Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.269832 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-api" containerID="cri-o://5047209590a7b4f61642164ff4a9927483d2a43c6caf68e8a2aa40527b28b88e" gracePeriod=30 Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.309692 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.309915 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" containerName="nova-scheduler-scheduler" containerID="cri-o://48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474" gracePeriod=30 Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.332902 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74f3d684-e75b-4b86-9086-e2649c75ee9e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.338748 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.339184 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-metadata" containerID="cri-o://da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5" gracePeriod=30 Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.339046 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-log" containerID="cri-o://6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e" gracePeriod=30 Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.448463 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dfd9d95f-cvfsd"] Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.458626 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6dfd9d95f-cvfsd"] Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.724153 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:50:20 crc kubenswrapper[4750]: I0309 18:50:20.975237 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.130834 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.137580 4750 generic.go:334] "Generic (PLEG): container finished" podID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerID="1e7723029dd26e346b160c3e043bdae867f720f655aaa6001ee91e73b445d04b" exitCode=143 Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.137737 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e2a829f-ac14-4141-a741-9ad6c1615f52","Type":"ContainerDied","Data":"1e7723029dd26e346b160c3e043bdae867f720f655aaa6001ee91e73b445d04b"} Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.139063 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerStarted","Data":"93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6"} Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.139098 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerStarted","Data":"b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5"} Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.139113 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerStarted","Data":"6bdebef4f8e2e04117d1f54e84805a603727706491433e8aadb902b220b5e6f4"} Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.140514 4750 generic.go:334] "Generic (PLEG): container finished" podID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerID="da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5" exitCode=0 Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.140542 4750 generic.go:334] "Generic (PLEG): container finished" podID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerID="6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e" exitCode=143 Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.140727 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"498246e2-9cdd-4a11-984b-3b703f2108bc","Type":"ContainerDied","Data":"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5"} Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.140782 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"498246e2-9cdd-4a11-984b-3b703f2108bc","Type":"ContainerDied","Data":"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e"} Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.140796 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"498246e2-9cdd-4a11-984b-3b703f2108bc","Type":"ContainerDied","Data":"8f4d6a490026e437d49729ff8b369d2890cac5ee22ae5d33872436c78be988e3"} Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.140814 4750 scope.go:117] "RemoveContainer" containerID="da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.140938 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.210585 4750 scope.go:117] "RemoveContainer" containerID="6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.239284 4750 scope.go:117] "RemoveContainer" containerID="da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5" Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.241840 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5\": container with ID starting with da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5 not found: ID does not exist" containerID="da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.241886 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5"} err="failed to get container status \"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5\": rpc error: code = NotFound desc = could not find container \"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5\": container with ID starting with da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5 not found: ID does not exist" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.241914 4750 scope.go:117] "RemoveContainer" containerID="6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e" Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.242380 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e\": container with ID starting with 6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e not found: ID does not exist" containerID="6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.242403 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e"} err="failed to get container status \"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e\": rpc error: code = NotFound desc = could not find container \"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e\": container with ID starting with 6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e not found: ID does not exist" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.242420 4750 scope.go:117] "RemoveContainer" containerID="da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.243145 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5"} err="failed to get container status \"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5\": rpc error: code = NotFound desc = could not find container \"da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5\": container with ID starting with da4f8bd7711ce4a005545377bae7ebaa95f2a2b3d1aeb35ae7b24455283c58a5 not found: ID does not exist" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.243163 4750 scope.go:117] "RemoveContainer" containerID="6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.243365 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e"} err="failed to get container status \"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e\": rpc error: code = NotFound desc = could not find container \"6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e\": container with ID starting with 6f3610e0628eb797f3a0fad8c5459b1f7b025dd59432c60d23c4a89671be6e5e not found: ID does not exist" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.255860 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498246e2-9cdd-4a11-984b-3b703f2108bc-logs\") pod \"498246e2-9cdd-4a11-984b-3b703f2108bc\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.255940 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f55nk\" (UniqueName: \"kubernetes.io/projected/498246e2-9cdd-4a11-984b-3b703f2108bc-kube-api-access-f55nk\") pod \"498246e2-9cdd-4a11-984b-3b703f2108bc\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.256070 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-combined-ca-bundle\") pod \"498246e2-9cdd-4a11-984b-3b703f2108bc\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.256291 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-nova-metadata-tls-certs\") pod \"498246e2-9cdd-4a11-984b-3b703f2108bc\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.256691 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-config-data\") pod \"498246e2-9cdd-4a11-984b-3b703f2108bc\" (UID: \"498246e2-9cdd-4a11-984b-3b703f2108bc\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.257082 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/498246e2-9cdd-4a11-984b-3b703f2108bc-logs" (OuterVolumeSpecName: "logs") pod "498246e2-9cdd-4a11-984b-3b703f2108bc" (UID: "498246e2-9cdd-4a11-984b-3b703f2108bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.258505 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/498246e2-9cdd-4a11-984b-3b703f2108bc-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.262847 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/498246e2-9cdd-4a11-984b-3b703f2108bc-kube-api-access-f55nk" (OuterVolumeSpecName: "kube-api-access-f55nk") pod "498246e2-9cdd-4a11-984b-3b703f2108bc" (UID: "498246e2-9cdd-4a11-984b-3b703f2108bc"). InnerVolumeSpecName "kube-api-access-f55nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.304824 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-config-data" (OuterVolumeSpecName: "config-data") pod "498246e2-9cdd-4a11-984b-3b703f2108bc" (UID: "498246e2-9cdd-4a11-984b-3b703f2108bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.308851 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "498246e2-9cdd-4a11-984b-3b703f2108bc" (UID: "498246e2-9cdd-4a11-984b-3b703f2108bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.337802 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "498246e2-9cdd-4a11-984b-3b703f2108bc" (UID: "498246e2-9cdd-4a11-984b-3b703f2108bc"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.360892 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f55nk\" (UniqueName: \"kubernetes.io/projected/498246e2-9cdd-4a11-984b-3b703f2108bc-kube-api-access-f55nk\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.360931 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.360939 4750 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.360948 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498246e2-9cdd-4a11-984b-3b703f2108bc-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.397794 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74f3d684-e75b-4b86-9086-e2649c75ee9e" path="/var/lib/kubelet/pods/74f3d684-e75b-4b86-9086-e2649c75ee9e/volumes" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.476772 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.489186 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.498649 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.513699 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fsdkk"] Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.546592 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.547062 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" containerName="nova-cell1-conductor-db-sync" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547079 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" containerName="nova-cell1-conductor-db-sync" Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.547094 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-log" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547100 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-log" Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.547112 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerName="dnsmasq-dns" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547119 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerName="dnsmasq-dns" Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.547135 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df584b8-cb6c-4044-a738-8500b2a3e78b" containerName="nova-manage" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547141 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df584b8-cb6c-4044-a738-8500b2a3e78b" containerName="nova-manage" Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.547159 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerName="init" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547165 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerName="init" Mar 09 18:50:21 crc kubenswrapper[4750]: E0309 18:50:21.547178 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-metadata" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547183 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-metadata" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547365 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df584b8-cb6c-4044-a738-8500b2a3e78b" containerName="nova-manage" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547381 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-metadata" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547393 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" containerName="nova-metadata-log" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547402 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" containerName="nova-cell1-conductor-db-sync" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.547409 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="74f3d684-e75b-4b86-9086-e2649c75ee9e" containerName="dnsmasq-dns" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.548468 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.550739 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.551999 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.563554 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-597mq\" (UniqueName: \"kubernetes.io/projected/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-kube-api-access-597mq\") pod \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.563596 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-scripts\") pod \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.563815 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-config-data\") pod \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.563858 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-combined-ca-bundle\") pod \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\" (UID: \"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0\") " Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.566463 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.574886 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-scripts" (OuterVolumeSpecName: "scripts") pod "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" (UID: "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.575624 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-kube-api-access-597mq" (OuterVolumeSpecName: "kube-api-access-597mq") pod "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" (UID: "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0"). InnerVolumeSpecName "kube-api-access-597mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.602283 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" (UID: "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.623717 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-config-data" (OuterVolumeSpecName: "config-data") pod "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" (UID: "6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.668622 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26j9x\" (UniqueName: \"kubernetes.io/projected/2359c431-502c-4c73-8618-d871ec51bfc9-kube-api-access-26j9x\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.668855 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2359c431-502c-4c73-8618-d871ec51bfc9-logs\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.669228 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.678159 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.678387 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-config-data\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.678551 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-597mq\" (UniqueName: \"kubernetes.io/projected/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-kube-api-access-597mq\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.678564 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.678576 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.678585 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.743944 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.744009 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.780097 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26j9x\" (UniqueName: \"kubernetes.io/projected/2359c431-502c-4c73-8618-d871ec51bfc9-kube-api-access-26j9x\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.780151 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2359c431-502c-4c73-8618-d871ec51bfc9-logs\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.780237 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.780282 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.780331 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-config-data\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.780764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2359c431-502c-4c73-8618-d871ec51bfc9-logs\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.784055 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.784506 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-config-data\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.790298 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.799896 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26j9x\" (UniqueName: \"kubernetes.io/projected/2359c431-502c-4c73-8618-d871ec51bfc9-kube-api-access-26j9x\") pod \"nova-metadata-0\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " pod="openstack/nova-metadata-0" Mar 09 18:50:21 crc kubenswrapper[4750]: I0309 18:50:21.873411 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.200939 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.202972 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.204220 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l99gb" event={"ID":"6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0","Type":"ContainerDied","Data":"c8932a448a06807f40d3c4df7f8838c923d1f6c47dabb9116bdcb071c666c011"} Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.204250 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8932a448a06807f40d3c4df7f8838c923d1f6c47dabb9116bdcb071c666c011" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.204306 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l99gb" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.233677 4750 generic.go:334] "Generic (PLEG): container finished" podID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerID="5047209590a7b4f61642164ff4a9927483d2a43c6caf68e8a2aa40527b28b88e" exitCode=0 Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.233741 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e2a829f-ac14-4141-a741-9ad6c1615f52","Type":"ContainerDied","Data":"5047209590a7b4f61642164ff4a9927483d2a43c6caf68e8a2aa40527b28b88e"} Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.234410 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.246005 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerStarted","Data":"5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f"} Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.251231 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fsdkk" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="registry-server" containerID="cri-o://85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae" gracePeriod=2 Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.291043 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78014cb6-5fde-4f65-990e-9c46da7a8578-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.291086 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqcpz\" (UniqueName: \"kubernetes.io/projected/78014cb6-5fde-4f65-990e-9c46da7a8578-kube-api-access-gqcpz\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.291196 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78014cb6-5fde-4f65-990e-9c46da7a8578-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.393497 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78014cb6-5fde-4f65-990e-9c46da7a8578-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.393531 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqcpz\" (UniqueName: \"kubernetes.io/projected/78014cb6-5fde-4f65-990e-9c46da7a8578-kube-api-access-gqcpz\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.393594 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78014cb6-5fde-4f65-990e-9c46da7a8578-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.401497 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78014cb6-5fde-4f65-990e-9c46da7a8578-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.403913 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78014cb6-5fde-4f65-990e-9c46da7a8578-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.418558 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqcpz\" (UniqueName: \"kubernetes.io/projected/78014cb6-5fde-4f65-990e-9c46da7a8578-kube-api-access-gqcpz\") pod \"nova-cell1-conductor-0\" (UID: \"78014cb6-5fde-4f65-990e-9c46da7a8578\") " pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.494591 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.517468 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.541354 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.549153 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.597327 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk2lv\" (UniqueName: \"kubernetes.io/projected/9e2a829f-ac14-4141-a741-9ad6c1615f52-kube-api-access-qk2lv\") pod \"9e2a829f-ac14-4141-a741-9ad6c1615f52\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.597396 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2a829f-ac14-4141-a741-9ad6c1615f52-logs\") pod \"9e2a829f-ac14-4141-a741-9ad6c1615f52\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.597567 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-config-data\") pod \"9e2a829f-ac14-4141-a741-9ad6c1615f52\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.597663 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-combined-ca-bundle\") pod \"9e2a829f-ac14-4141-a741-9ad6c1615f52\" (UID: \"9e2a829f-ac14-4141-a741-9ad6c1615f52\") " Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.598038 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e2a829f-ac14-4141-a741-9ad6c1615f52-logs" (OuterVolumeSpecName: "logs") pod "9e2a829f-ac14-4141-a741-9ad6c1615f52" (UID: "9e2a829f-ac14-4141-a741-9ad6c1615f52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.598228 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2a829f-ac14-4141-a741-9ad6c1615f52-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.611918 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2a829f-ac14-4141-a741-9ad6c1615f52-kube-api-access-qk2lv" (OuterVolumeSpecName: "kube-api-access-qk2lv") pod "9e2a829f-ac14-4141-a741-9ad6c1615f52" (UID: "9e2a829f-ac14-4141-a741-9ad6c1615f52"). InnerVolumeSpecName "kube-api-access-qk2lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.636204 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-config-data" (OuterVolumeSpecName: "config-data") pod "9e2a829f-ac14-4141-a741-9ad6c1615f52" (UID: "9e2a829f-ac14-4141-a741-9ad6c1615f52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.648168 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e2a829f-ac14-4141-a741-9ad6c1615f52" (UID: "9e2a829f-ac14-4141-a741-9ad6c1615f52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.729016 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk2lv\" (UniqueName: \"kubernetes.io/projected/9e2a829f-ac14-4141-a741-9ad6c1615f52-kube-api-access-qk2lv\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.729053 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.729064 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2a829f-ac14-4141-a741-9ad6c1615f52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.756848 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.830044 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tsp4\" (UniqueName: \"kubernetes.io/projected/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-kube-api-access-5tsp4\") pod \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.830120 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-utilities\") pod \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.830187 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-catalog-content\") pod \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\" (UID: \"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2\") " Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.831362 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-utilities" (OuterVolumeSpecName: "utilities") pod "4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" (UID: "4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.842699 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-kube-api-access-5tsp4" (OuterVolumeSpecName: "kube-api-access-5tsp4") pod "4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" (UID: "4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2"). InnerVolumeSpecName "kube-api-access-5tsp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.932016 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tsp4\" (UniqueName: \"kubernetes.io/projected/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-kube-api-access-5tsp4\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.932061 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:22 crc kubenswrapper[4750]: I0309 18:50:22.959867 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" (UID: "4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.033964 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.194570 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.271399 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2359c431-502c-4c73-8618-d871ec51bfc9","Type":"ContainerStarted","Data":"f380d429b5b30034c5ab645d5a7d07a77b9959ffdd9b3c9d697d611ce1a7672f"} Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.271447 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2359c431-502c-4c73-8618-d871ec51bfc9","Type":"ContainerStarted","Data":"56085c5c1f7999df142c9835426770bc49dc1dd4df13c18e365dc9d4e4662192"} Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.271461 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2359c431-502c-4c73-8618-d871ec51bfc9","Type":"ContainerStarted","Data":"04f0fbda3a91280de13e5bafc5a15c80964dda3c5facbdea8e32cdb16051ea62"} Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.276060 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e2a829f-ac14-4141-a741-9ad6c1615f52","Type":"ContainerDied","Data":"fc326e617fb79fcf0e6fc4c7c2dcaa6bc0182377376c34ed28b0a15285fe256e"} Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.276095 4750 scope.go:117] "RemoveContainer" containerID="5047209590a7b4f61642164ff4a9927483d2a43c6caf68e8a2aa40527b28b88e" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.276117 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.280121 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerID="85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae" exitCode=0 Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.280156 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fsdkk" event={"ID":"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2","Type":"ContainerDied","Data":"85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae"} Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.280193 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fsdkk" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.280200 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fsdkk" event={"ID":"4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2","Type":"ContainerDied","Data":"e24f4ac88d68acc9105bed79fe01177e71c281cb3623efd53b4b1f0dde8d6bcf"} Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.281866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"78014cb6-5fde-4f65-990e-9c46da7a8578","Type":"ContainerStarted","Data":"b4d577d63d2b74edc9658a7d17637ca04f10d00249e70414018f3a6f43e6f73a"} Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.300644 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.3005931840000002 podStartE2EDuration="2.300593184s" podCreationTimestamp="2026-03-09 18:50:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:23.286469729 +0000 UTC m=+1504.628942127" watchObservedRunningTime="2026-03-09 18:50:23.300593184 +0000 UTC m=+1504.643065582" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.323545 4750 scope.go:117] "RemoveContainer" containerID="1e7723029dd26e346b160c3e043bdae867f720f655aaa6001ee91e73b445d04b" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.323690 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fsdkk"] Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.334897 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fsdkk"] Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.353070 4750 scope.go:117] "RemoveContainer" containerID="85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.371298 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.391211 4750 scope.go:117] "RemoveContainer" containerID="2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.403583 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="498246e2-9cdd-4a11-984b-3b703f2108bc" path="/var/lib/kubelet/pods/498246e2-9cdd-4a11-984b-3b703f2108bc/volumes" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.404478 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" path="/var/lib/kubelet/pods/4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2/volumes" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.407044 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.425769 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.426186 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-log" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426204 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-log" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.426230 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="registry-server" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426237 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="registry-server" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.426249 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="extract-content" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426255 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="extract-content" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.426265 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-api" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426271 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-api" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.426278 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="extract-utilities" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426284 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="extract-utilities" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426452 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-log" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426467 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" containerName="nova-api-api" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.426477 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d181d52-b8aa-4dc2-8fd8-38fc4a90b6f2" containerName="registry-server" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.427496 4750 scope.go:117] "RemoveContainer" containerID="7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.427669 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.430839 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.443802 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.491908 4750 scope.go:117] "RemoveContainer" containerID="85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.493193 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae\": container with ID starting with 85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae not found: ID does not exist" containerID="85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.493281 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae"} err="failed to get container status \"85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae\": rpc error: code = NotFound desc = could not find container \"85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae\": container with ID starting with 85533c431e1b43a3d0a5f44f904c8e4441ce7a00e9949cb1a45e0619e967c4ae not found: ID does not exist" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.493327 4750 scope.go:117] "RemoveContainer" containerID="2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.495085 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653\": container with ID starting with 2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653 not found: ID does not exist" containerID="2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.495118 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653"} err="failed to get container status \"2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653\": rpc error: code = NotFound desc = could not find container \"2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653\": container with ID starting with 2d20e4fe6e368e776d51a837c4d22995bf8090b7c6a12c4d2404da90a1255653 not found: ID does not exist" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.495134 4750 scope.go:117] "RemoveContainer" containerID="7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.495479 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e\": container with ID starting with 7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e not found: ID does not exist" containerID="7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.495524 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e"} err="failed to get container status \"7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e\": rpc error: code = NotFound desc = could not find container \"7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e\": container with ID starting with 7fc9fd043973ab7966702470b7798b8f04d65936e1d66d9abd49f3f9ead9ab6e not found: ID does not exist" Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.502963 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.508894 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.515034 4750 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 09 18:50:23 crc kubenswrapper[4750]: E0309 18:50:23.515214 4750 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" containerName="nova-scheduler-scheduler" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.547615 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.547787 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9g44\" (UniqueName: \"kubernetes.io/projected/8c2b6a83-2323-4271-a679-85b9c533c5cc-kube-api-access-r9g44\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.547829 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-config-data\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.548290 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2b6a83-2323-4271-a679-85b9c533c5cc-logs\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.649793 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9g44\" (UniqueName: \"kubernetes.io/projected/8c2b6a83-2323-4271-a679-85b9c533c5cc-kube-api-access-r9g44\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.649850 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-config-data\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.649951 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2b6a83-2323-4271-a679-85b9c533c5cc-logs\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.649997 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.650500 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2b6a83-2323-4271-a679-85b9c533c5cc-logs\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.655570 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-config-data\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.663299 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.666335 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9g44\" (UniqueName: \"kubernetes.io/projected/8c2b6a83-2323-4271-a679-85b9c533c5cc-kube-api-access-r9g44\") pod \"nova-api-0\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " pod="openstack/nova-api-0" Mar 09 18:50:23 crc kubenswrapper[4750]: I0309 18:50:23.752798 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:24 crc kubenswrapper[4750]: I0309 18:50:24.274529 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:24 crc kubenswrapper[4750]: I0309 18:50:24.294477 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c2b6a83-2323-4271-a679-85b9c533c5cc","Type":"ContainerStarted","Data":"a63b77d45f8f428e7c15a8a6621ce4b92c598dee56560ffd9552a6f4ac03fe6e"} Mar 09 18:50:24 crc kubenswrapper[4750]: I0309 18:50:24.298157 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"78014cb6-5fde-4f65-990e-9c46da7a8578","Type":"ContainerStarted","Data":"186c03458ddbc52f643ca6cc270d09915ab1b5a2534f1e96c6e363c894c7ac03"} Mar 09 18:50:24 crc kubenswrapper[4750]: I0309 18:50:24.299543 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:24 crc kubenswrapper[4750]: I0309 18:50:24.348409 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.348385382 podStartE2EDuration="2.348385382s" podCreationTimestamp="2026-03-09 18:50:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:24.336723114 +0000 UTC m=+1505.679195512" watchObservedRunningTime="2026-03-09 18:50:24.348385382 +0000 UTC m=+1505.690857800" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.351578 4750 generic.go:334] "Generic (PLEG): container finished" podID="7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" containerID="48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474" exitCode=0 Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.352257 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9","Type":"ContainerDied","Data":"48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474"} Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.362997 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c2b6a83-2323-4271-a679-85b9c533c5cc","Type":"ContainerStarted","Data":"cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528"} Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.363044 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c2b6a83-2323-4271-a679-85b9c533c5cc","Type":"ContainerStarted","Data":"0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047"} Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.367321 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerStarted","Data":"fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759"} Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.367472 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.399498 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.39947987 podStartE2EDuration="2.39947987s" podCreationTimestamp="2026-03-09 18:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:25.388542862 +0000 UTC m=+1506.731015260" watchObservedRunningTime="2026-03-09 18:50:25.39947987 +0000 UTC m=+1506.741952268" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.399565 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e2a829f-ac14-4141-a741-9ad6c1615f52" path="/var/lib/kubelet/pods/9e2a829f-ac14-4141-a741-9ad6c1615f52/volumes" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.417413 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.717768486 podStartE2EDuration="6.417392639s" podCreationTimestamp="2026-03-09 18:50:19 +0000 UTC" firstStartedPulling="2026-03-09 18:50:20.166162332 +0000 UTC m=+1501.508634730" lastFinishedPulling="2026-03-09 18:50:24.865786495 +0000 UTC m=+1506.208258883" observedRunningTime="2026-03-09 18:50:25.413684447 +0000 UTC m=+1506.756156855" watchObservedRunningTime="2026-03-09 18:50:25.417392639 +0000 UTC m=+1506.759865037" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.453230 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.491291 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-combined-ca-bundle\") pod \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.491355 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-config-data\") pod \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.491398 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjg9h\" (UniqueName: \"kubernetes.io/projected/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-kube-api-access-tjg9h\") pod \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\" (UID: \"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9\") " Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.502840 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-kube-api-access-tjg9h" (OuterVolumeSpecName: "kube-api-access-tjg9h") pod "7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" (UID: "7827b0ea-5a6a-4c87-90ba-a75ae76d64a9"). InnerVolumeSpecName "kube-api-access-tjg9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.550842 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-config-data" (OuterVolumeSpecName: "config-data") pod "7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" (UID: "7827b0ea-5a6a-4c87-90ba-a75ae76d64a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.556666 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" (UID: "7827b0ea-5a6a-4c87-90ba-a75ae76d64a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.594432 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.594465 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:25 crc kubenswrapper[4750]: I0309 18:50:25.594474 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjg9h\" (UniqueName: \"kubernetes.io/projected/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9-kube-api-access-tjg9h\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.389042 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7827b0ea-5a6a-4c87-90ba-a75ae76d64a9","Type":"ContainerDied","Data":"badd727d8cbfbd2ed0f63bab07fff118fb3577ed90b042164dd07aefd0ad4a27"} Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.390877 4750 scope.go:117] "RemoveContainer" containerID="48de5574b0fd95680fc2c7a85e3240e8ca6db285f6cfd63e5956e3acfa700474" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.389135 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.447170 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.471785 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.482513 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:26 crc kubenswrapper[4750]: E0309 18:50:26.483049 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" containerName="nova-scheduler-scheduler" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.483071 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" containerName="nova-scheduler-scheduler" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.483333 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" containerName="nova-scheduler-scheduler" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.484183 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.492268 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.513432 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7klpv\" (UniqueName: \"kubernetes.io/projected/b2b28e7a-8d56-469f-bbd3-25524fb5a438-kube-api-access-7klpv\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.513524 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-config-data\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.513707 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.515265 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.616921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.617459 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7klpv\" (UniqueName: \"kubernetes.io/projected/b2b28e7a-8d56-469f-bbd3-25524fb5a438-kube-api-access-7klpv\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.617578 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-config-data\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.624220 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-config-data\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.624235 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.641670 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7klpv\" (UniqueName: \"kubernetes.io/projected/b2b28e7a-8d56-469f-bbd3-25524fb5a438-kube-api-access-7klpv\") pod \"nova-scheduler-0\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.807923 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.873516 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 09 18:50:26 crc kubenswrapper[4750]: I0309 18:50:26.873730 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 09 18:50:27 crc kubenswrapper[4750]: I0309 18:50:27.307249 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:50:27 crc kubenswrapper[4750]: I0309 18:50:27.386097 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7827b0ea-5a6a-4c87-90ba-a75ae76d64a9" path="/var/lib/kubelet/pods/7827b0ea-5a6a-4c87-90ba-a75ae76d64a9/volumes" Mar 09 18:50:27 crc kubenswrapper[4750]: I0309 18:50:27.398710 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2b28e7a-8d56-469f-bbd3-25524fb5a438","Type":"ContainerStarted","Data":"5470a992ec245a9c25a8a2b2b46d69dc68d92b9f5ed763d78f73ea3de4dee81c"} Mar 09 18:50:28 crc kubenswrapper[4750]: I0309 18:50:28.412556 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2b28e7a-8d56-469f-bbd3-25524fb5a438","Type":"ContainerStarted","Data":"ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6"} Mar 09 18:50:28 crc kubenswrapper[4750]: I0309 18:50:28.431328 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.431303937 podStartE2EDuration="2.431303937s" podCreationTimestamp="2026-03-09 18:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:28.427685559 +0000 UTC m=+1509.770157967" watchObservedRunningTime="2026-03-09 18:50:28.431303937 +0000 UTC m=+1509.773776365" Mar 09 18:50:31 crc kubenswrapper[4750]: I0309 18:50:31.808611 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 09 18:50:31 crc kubenswrapper[4750]: I0309 18:50:31.873715 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 09 18:50:31 crc kubenswrapper[4750]: I0309 18:50:31.875666 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 09 18:50:32 crc kubenswrapper[4750]: I0309 18:50:32.571346 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 09 18:50:32 crc kubenswrapper[4750]: I0309 18:50:32.893025 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 09 18:50:32 crc kubenswrapper[4750]: I0309 18:50:32.893027 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 09 18:50:33 crc kubenswrapper[4750]: I0309 18:50:33.754183 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:50:33 crc kubenswrapper[4750]: I0309 18:50:33.754272 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:50:34 crc kubenswrapper[4750]: I0309 18:50:34.836881 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.232:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 09 18:50:34 crc kubenswrapper[4750]: I0309 18:50:34.836881 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.232:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 09 18:50:36 crc kubenswrapper[4750]: I0309 18:50:36.808756 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 09 18:50:36 crc kubenswrapper[4750]: I0309 18:50:36.843120 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 09 18:50:37 crc kubenswrapper[4750]: I0309 18:50:37.548056 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 09 18:50:40 crc kubenswrapper[4750]: I0309 18:50:40.984856 4750 scope.go:117] "RemoveContainer" containerID="a1e7416aecca045ff84400c55306c667a0c4ff18352fdd539eb0c924c2c24b47" Mar 09 18:50:41 crc kubenswrapper[4750]: I0309 18:50:41.882122 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 09 18:50:41 crc kubenswrapper[4750]: I0309 18:50:41.882362 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 09 18:50:41 crc kubenswrapper[4750]: I0309 18:50:41.891290 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 09 18:50:41 crc kubenswrapper[4750]: I0309 18:50:41.893589 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.391826 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.513779 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-combined-ca-bundle\") pod \"f746758c-2dda-48c8-aead-fad027da9e6e\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.513893 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2srzz\" (UniqueName: \"kubernetes.io/projected/f746758c-2dda-48c8-aead-fad027da9e6e-kube-api-access-2srzz\") pod \"f746758c-2dda-48c8-aead-fad027da9e6e\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.514748 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-config-data\") pod \"f746758c-2dda-48c8-aead-fad027da9e6e\" (UID: \"f746758c-2dda-48c8-aead-fad027da9e6e\") " Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.530456 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f746758c-2dda-48c8-aead-fad027da9e6e-kube-api-access-2srzz" (OuterVolumeSpecName: "kube-api-access-2srzz") pod "f746758c-2dda-48c8-aead-fad027da9e6e" (UID: "f746758c-2dda-48c8-aead-fad027da9e6e"). InnerVolumeSpecName "kube-api-access-2srzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.544980 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f746758c-2dda-48c8-aead-fad027da9e6e" (UID: "f746758c-2dda-48c8-aead-fad027da9e6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.555972 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-config-data" (OuterVolumeSpecName: "config-data") pod "f746758c-2dda-48c8-aead-fad027da9e6e" (UID: "f746758c-2dda-48c8-aead-fad027da9e6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.584126 4750 generic.go:334] "Generic (PLEG): container finished" podID="f746758c-2dda-48c8-aead-fad027da9e6e" containerID="7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4" exitCode=137 Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.584222 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.584256 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f746758c-2dda-48c8-aead-fad027da9e6e","Type":"ContainerDied","Data":"7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4"} Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.584311 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f746758c-2dda-48c8-aead-fad027da9e6e","Type":"ContainerDied","Data":"b6cfa6e1727c2ed1ddb52d0bcbc8c61c294ea8f91164bd2ac1b35d3802777d44"} Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.584335 4750 scope.go:117] "RemoveContainer" containerID="7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.622501 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.622546 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2srzz\" (UniqueName: \"kubernetes.io/projected/f746758c-2dda-48c8-aead-fad027da9e6e-kube-api-access-2srzz\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.622563 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f746758c-2dda-48c8-aead-fad027da9e6e-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.637148 4750 scope.go:117] "RemoveContainer" containerID="7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4" Mar 09 18:50:42 crc kubenswrapper[4750]: E0309 18:50:42.637685 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4\": container with ID starting with 7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4 not found: ID does not exist" containerID="7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.637730 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4"} err="failed to get container status \"7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4\": rpc error: code = NotFound desc = could not find container \"7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4\": container with ID starting with 7673689299b407e9fce396bb8eb5ca56516464b9ea0d69d5afaa0904235736b4 not found: ID does not exist" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.640500 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.650971 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.671443 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:42 crc kubenswrapper[4750]: E0309 18:50:42.672055 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f746758c-2dda-48c8-aead-fad027da9e6e" containerName="nova-cell1-novncproxy-novncproxy" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.672073 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f746758c-2dda-48c8-aead-fad027da9e6e" containerName="nova-cell1-novncproxy-novncproxy" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.672322 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f746758c-2dda-48c8-aead-fad027da9e6e" containerName="nova-cell1-novncproxy-novncproxy" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.673091 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.677562 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.677577 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.678659 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.683092 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.831388 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.831490 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.831616 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.831652 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdz8w\" (UniqueName: \"kubernetes.io/projected/bfba6047-48a3-4895-90c1-fadceadccba3-kube-api-access-jdz8w\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.831679 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.933892 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.933953 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdz8w\" (UniqueName: \"kubernetes.io/projected/bfba6047-48a3-4895-90c1-fadceadccba3-kube-api-access-jdz8w\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.933992 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.934083 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.934165 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.939574 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.939834 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.940028 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.941018 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfba6047-48a3-4895-90c1-fadceadccba3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.960799 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdz8w\" (UniqueName: \"kubernetes.io/projected/bfba6047-48a3-4895-90c1-fadceadccba3-kube-api-access-jdz8w\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfba6047-48a3-4895-90c1-fadceadccba3\") " pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:42 crc kubenswrapper[4750]: I0309 18:50:42.988460 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:43 crc kubenswrapper[4750]: I0309 18:50:43.385779 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f746758c-2dda-48c8-aead-fad027da9e6e" path="/var/lib/kubelet/pods/f746758c-2dda-48c8-aead-fad027da9e6e/volumes" Mar 09 18:50:43 crc kubenswrapper[4750]: I0309 18:50:43.506285 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 09 18:50:43 crc kubenswrapper[4750]: W0309 18:50:43.528416 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfba6047_48a3_4895_90c1_fadceadccba3.slice/crio-003980b497994a57193408b74a2621e00c40eaa24a394b7b25aa1d44362c4f21 WatchSource:0}: Error finding container 003980b497994a57193408b74a2621e00c40eaa24a394b7b25aa1d44362c4f21: Status 404 returned error can't find the container with id 003980b497994a57193408b74a2621e00c40eaa24a394b7b25aa1d44362c4f21 Mar 09 18:50:43 crc kubenswrapper[4750]: I0309 18:50:43.598582 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bfba6047-48a3-4895-90c1-fadceadccba3","Type":"ContainerStarted","Data":"003980b497994a57193408b74a2621e00c40eaa24a394b7b25aa1d44362c4f21"} Mar 09 18:50:43 crc kubenswrapper[4750]: I0309 18:50:43.761211 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 09 18:50:43 crc kubenswrapper[4750]: I0309 18:50:43.762386 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 09 18:50:43 crc kubenswrapper[4750]: I0309 18:50:43.763824 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 09 18:50:43 crc kubenswrapper[4750]: I0309 18:50:43.787809 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.613855 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bfba6047-48a3-4895-90c1-fadceadccba3","Type":"ContainerStarted","Data":"82602f058e0abdfc3aaaa61ea262cd9791653850a209625e4b42501035d54c28"} Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.614666 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.621842 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.638422 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.638389427 podStartE2EDuration="2.638389427s" podCreationTimestamp="2026-03-09 18:50:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:44.637094873 +0000 UTC m=+1525.979567271" watchObservedRunningTime="2026-03-09 18:50:44.638389427 +0000 UTC m=+1525.980861865" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.827837 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89669fc97-cr5m9"] Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.829514 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.860016 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89669fc97-cr5m9"] Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.881349 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-svc\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.881434 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6wbg\" (UniqueName: \"kubernetes.io/projected/46f02dee-def0-448a-8813-cf4e9083fec2-kube-api-access-z6wbg\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.881467 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-swift-storage-0\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.881519 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-sb\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.881542 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-nb\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.881617 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-config\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.987401 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-sb\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.987444 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-nb\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.987515 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-config\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.987597 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-svc\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.987651 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6wbg\" (UniqueName: \"kubernetes.io/projected/46f02dee-def0-448a-8813-cf4e9083fec2-kube-api-access-z6wbg\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.987675 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-swift-storage-0\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.988830 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-nb\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.988893 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-swift-storage-0\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.989447 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-config\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.990595 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-sb\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:44 crc kubenswrapper[4750]: I0309 18:50:44.993428 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-svc\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:45 crc kubenswrapper[4750]: I0309 18:50:45.020778 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6wbg\" (UniqueName: \"kubernetes.io/projected/46f02dee-def0-448a-8813-cf4e9083fec2-kube-api-access-z6wbg\") pod \"dnsmasq-dns-89669fc97-cr5m9\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:45 crc kubenswrapper[4750]: I0309 18:50:45.162230 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:45 crc kubenswrapper[4750]: I0309 18:50:45.847708 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89669fc97-cr5m9"] Mar 09 18:50:45 crc kubenswrapper[4750]: W0309 18:50:45.851784 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46f02dee_def0_448a_8813_cf4e9083fec2.slice/crio-8fec15ec5a11ca324f9f2d6aeedcc4013248ae4d53de78e4bab9c42b45271497 WatchSource:0}: Error finding container 8fec15ec5a11ca324f9f2d6aeedcc4013248ae4d53de78e4bab9c42b45271497: Status 404 returned error can't find the container with id 8fec15ec5a11ca324f9f2d6aeedcc4013248ae4d53de78e4bab9c42b45271497 Mar 09 18:50:46 crc kubenswrapper[4750]: I0309 18:50:46.636208 4750 generic.go:334] "Generic (PLEG): container finished" podID="46f02dee-def0-448a-8813-cf4e9083fec2" containerID="e87ffa1b9c2af86bc4fcc1571081142655397fce9bae64cdda8a84223b111c11" exitCode=0 Mar 09 18:50:46 crc kubenswrapper[4750]: I0309 18:50:46.636273 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" event={"ID":"46f02dee-def0-448a-8813-cf4e9083fec2","Type":"ContainerDied","Data":"e87ffa1b9c2af86bc4fcc1571081142655397fce9bae64cdda8a84223b111c11"} Mar 09 18:50:46 crc kubenswrapper[4750]: I0309 18:50:46.636596 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" event={"ID":"46f02dee-def0-448a-8813-cf4e9083fec2","Type":"ContainerStarted","Data":"8fec15ec5a11ca324f9f2d6aeedcc4013248ae4d53de78e4bab9c42b45271497"} Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.289576 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.290473 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-central-agent" containerID="cri-o://b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5" gracePeriod=30 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.290621 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="proxy-httpd" containerID="cri-o://fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759" gracePeriod=30 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.290711 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-notification-agent" containerID="cri-o://93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6" gracePeriod=30 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.290622 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="sg-core" containerID="cri-o://5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f" gracePeriod=30 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.300559 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.229:3000/\": EOF" Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.446242 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.650114 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" event={"ID":"46f02dee-def0-448a-8813-cf4e9083fec2","Type":"ContainerStarted","Data":"bf7e8ab45df770e7d117f6e00b482d9df09dcc1557d1974dd5f480e7beda8062"} Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.650445 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.656445 4750 generic.go:334] "Generic (PLEG): container finished" podID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerID="fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759" exitCode=0 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.656488 4750 generic.go:334] "Generic (PLEG): container finished" podID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerID="5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f" exitCode=2 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.656502 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerDied","Data":"fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759"} Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.656568 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerDied","Data":"5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f"} Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.656774 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-log" containerID="cri-o://0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047" gracePeriod=30 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.656821 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-api" containerID="cri-o://cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528" gracePeriod=30 Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.691301 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" podStartSLOduration=3.6912762089999998 podStartE2EDuration="3.691276209s" podCreationTimestamp="2026-03-09 18:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:47.687512816 +0000 UTC m=+1529.029985214" watchObservedRunningTime="2026-03-09 18:50:47.691276209 +0000 UTC m=+1529.033748607" Mar 09 18:50:47 crc kubenswrapper[4750]: I0309 18:50:47.988828 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:48 crc kubenswrapper[4750]: I0309 18:50:48.669226 4750 generic.go:334] "Generic (PLEG): container finished" podID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerID="b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5" exitCode=0 Mar 09 18:50:48 crc kubenswrapper[4750]: I0309 18:50:48.669330 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerDied","Data":"b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5"} Mar 09 18:50:48 crc kubenswrapper[4750]: I0309 18:50:48.671944 4750 generic.go:334] "Generic (PLEG): container finished" podID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerID="0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047" exitCode=143 Mar 09 18:50:48 crc kubenswrapper[4750]: I0309 18:50:48.672004 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c2b6a83-2323-4271-a679-85b9c533c5cc","Type":"ContainerDied","Data":"0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047"} Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.447191 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510105 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-config-data\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510261 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-log-httpd\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510382 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-run-httpd\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510458 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-scripts\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510500 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-combined-ca-bundle\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510525 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-ceilometer-tls-certs\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510553 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qx5x\" (UniqueName: \"kubernetes.io/projected/221bc157-a4af-4e07-89fa-1bbd1dcd254c-kube-api-access-4qx5x\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.510587 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-sg-core-conf-yaml\") pod \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\" (UID: \"221bc157-a4af-4e07-89fa-1bbd1dcd254c\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.518325 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.518339 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.524962 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-scripts" (OuterVolumeSpecName: "scripts") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.530036 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/221bc157-a4af-4e07-89fa-1bbd1dcd254c-kube-api-access-4qx5x" (OuterVolumeSpecName: "kube-api-access-4qx5x") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "kube-api-access-4qx5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.567204 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.619320 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qx5x\" (UniqueName: \"kubernetes.io/projected/221bc157-a4af-4e07-89fa-1bbd1dcd254c-kube-api-access-4qx5x\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.619358 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.619375 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.619388 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/221bc157-a4af-4e07-89fa-1bbd1dcd254c-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.619400 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.622476 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.652844 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.665091 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-config-data" (OuterVolumeSpecName: "config-data") pod "221bc157-a4af-4e07-89fa-1bbd1dcd254c" (UID: "221bc157-a4af-4e07-89fa-1bbd1dcd254c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.673746 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.692468 4750 generic.go:334] "Generic (PLEG): container finished" podID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerID="cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528" exitCode=0 Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.692591 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.693116 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c2b6a83-2323-4271-a679-85b9c533c5cc","Type":"ContainerDied","Data":"cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528"} Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.693210 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c2b6a83-2323-4271-a679-85b9c533c5cc","Type":"ContainerDied","Data":"a63b77d45f8f428e7c15a8a6621ce4b92c598dee56560ffd9552a6f4ac03fe6e"} Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.693233 4750 scope.go:117] "RemoveContainer" containerID="cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.709325 4750 generic.go:334] "Generic (PLEG): container finished" podID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerID="93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6" exitCode=0 Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.709391 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerDied","Data":"93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6"} Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.709422 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"221bc157-a4af-4e07-89fa-1bbd1dcd254c","Type":"ContainerDied","Data":"6bdebef4f8e2e04117d1f54e84805a603727706491433e8aadb902b220b5e6f4"} Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.709501 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.720273 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9g44\" (UniqueName: \"kubernetes.io/projected/8c2b6a83-2323-4271-a679-85b9c533c5cc-kube-api-access-r9g44\") pod \"8c2b6a83-2323-4271-a679-85b9c533c5cc\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.720560 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-config-data\") pod \"8c2b6a83-2323-4271-a679-85b9c533c5cc\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.720619 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2b6a83-2323-4271-a679-85b9c533c5cc-logs\") pod \"8c2b6a83-2323-4271-a679-85b9c533c5cc\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.720780 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-combined-ca-bundle\") pod \"8c2b6a83-2323-4271-a679-85b9c533c5cc\" (UID: \"8c2b6a83-2323-4271-a679-85b9c533c5cc\") " Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.721259 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.721284 4750 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.721296 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221bc157-a4af-4e07-89fa-1bbd1dcd254c-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.721357 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c2b6a83-2323-4271-a679-85b9c533c5cc-logs" (OuterVolumeSpecName: "logs") pod "8c2b6a83-2323-4271-a679-85b9c533c5cc" (UID: "8c2b6a83-2323-4271-a679-85b9c533c5cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.724482 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c2b6a83-2323-4271-a679-85b9c533c5cc-kube-api-access-r9g44" (OuterVolumeSpecName: "kube-api-access-r9g44") pod "8c2b6a83-2323-4271-a679-85b9c533c5cc" (UID: "8c2b6a83-2323-4271-a679-85b9c533c5cc"). InnerVolumeSpecName "kube-api-access-r9g44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.757004 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c2b6a83-2323-4271-a679-85b9c533c5cc" (UID: "8c2b6a83-2323-4271-a679-85b9c533c5cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.759741 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-config-data" (OuterVolumeSpecName: "config-data") pod "8c2b6a83-2323-4271-a679-85b9c533c5cc" (UID: "8c2b6a83-2323-4271-a679-85b9c533c5cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.760708 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.766356 4750 scope.go:117] "RemoveContainer" containerID="0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.794375 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.805036 4750 scope.go:117] "RemoveContainer" containerID="cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.805598 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528\": container with ID starting with cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528 not found: ID does not exist" containerID="cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.805664 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528"} err="failed to get container status \"cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528\": rpc error: code = NotFound desc = could not find container \"cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528\": container with ID starting with cdd6990c4f918e9e6a5c34125b70956e016103ad3eaa85a213f4f046e5f1b528 not found: ID does not exist" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.805697 4750 scope.go:117] "RemoveContainer" containerID="0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.806079 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047\": container with ID starting with 0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047 not found: ID does not exist" containerID="0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.806114 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047"} err="failed to get container status \"0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047\": rpc error: code = NotFound desc = could not find container \"0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047\": container with ID starting with 0d5fda35887ee8ad77ec8f8821a49534628fe6807e3c66a24b881f12578ca047 not found: ID does not exist" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.806128 4750 scope.go:117] "RemoveContainer" containerID="fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.808924 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.809489 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="proxy-httpd" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.809514 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="proxy-httpd" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.809534 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-central-agent" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.809540 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-central-agent" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.809552 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-notification-agent" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.809559 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-notification-agent" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.809573 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="sg-core" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.809579 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="sg-core" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.809601 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-api" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.809607 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-api" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.809641 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-log" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.809650 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-log" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.810048 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="sg-core" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.810073 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-log" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.810120 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" containerName="nova-api-api" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.810131 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-notification-agent" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.810219 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="ceilometer-central-agent" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.810284 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" containerName="proxy-httpd" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.812472 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.815616 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.817492 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.817807 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.822916 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.822940 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2b6a83-2323-4271-a679-85b9c533c5cc-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.822950 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2b6a83-2323-4271-a679-85b9c533c5cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.822959 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9g44\" (UniqueName: \"kubernetes.io/projected/8c2b6a83-2323-4271-a679-85b9c533c5cc-kube-api-access-r9g44\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.825787 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.848439 4750 scope.go:117] "RemoveContainer" containerID="5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.881745 4750 scope.go:117] "RemoveContainer" containerID="93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.897665 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.898875 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-wpkqw log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="bb796735-ed58-4ff1-aa33-1a8b63bc4905" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.919666 4750 scope.go:117] "RemoveContainer" containerID="b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925123 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925190 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-config-data\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925225 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925254 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-scripts\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925279 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925319 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-run-httpd\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925354 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpkqw\" (UniqueName: \"kubernetes.io/projected/bb796735-ed58-4ff1-aa33-1a8b63bc4905-kube-api-access-wpkqw\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.925377 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-log-httpd\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.951037 4750 scope.go:117] "RemoveContainer" containerID="fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.957842 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759\": container with ID starting with fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759 not found: ID does not exist" containerID="fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.957952 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759"} err="failed to get container status \"fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759\": rpc error: code = NotFound desc = could not find container \"fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759\": container with ID starting with fc550ca021dc3527aeec9e7cec6424c0af0d78054974f636ddc540df8ee2f759 not found: ID does not exist" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.958010 4750 scope.go:117] "RemoveContainer" containerID="5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.958803 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f\": container with ID starting with 5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f not found: ID does not exist" containerID="5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.958927 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f"} err="failed to get container status \"5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f\": rpc error: code = NotFound desc = could not find container \"5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f\": container with ID starting with 5720d154197c4f5bf706e318426147a6184771d19e8259f398ed62a3b44eac1f not found: ID does not exist" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.959021 4750 scope.go:117] "RemoveContainer" containerID="93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.959590 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6\": container with ID starting with 93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6 not found: ID does not exist" containerID="93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.959671 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6"} err="failed to get container status \"93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6\": rpc error: code = NotFound desc = could not find container \"93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6\": container with ID starting with 93aae4feeab712bb6375ea9a812e57c65fd1a44918c86f5f7fc84b80c5dff9e6 not found: ID does not exist" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.959702 4750 scope.go:117] "RemoveContainer" containerID="b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5" Mar 09 18:50:49 crc kubenswrapper[4750]: E0309 18:50:49.960077 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5\": container with ID starting with b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5 not found: ID does not exist" containerID="b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5" Mar 09 18:50:49 crc kubenswrapper[4750]: I0309 18:50:49.960167 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5"} err="failed to get container status \"b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5\": rpc error: code = NotFound desc = could not find container \"b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5\": container with ID starting with b02de593723f62f3576d400970a638cf2390b498225101070e9cc5ab9031b5a5 not found: ID does not exist" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.026910 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpkqw\" (UniqueName: \"kubernetes.io/projected/bb796735-ed58-4ff1-aa33-1a8b63bc4905-kube-api-access-wpkqw\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.026971 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-log-httpd\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.027106 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.027161 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-config-data\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.027186 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.027213 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-scripts\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.027230 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.027269 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-run-httpd\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.027460 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-log-httpd\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.028224 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-run-httpd\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.041295 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-config-data\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.044827 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.045400 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.047665 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-scripts\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.055319 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.062472 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpkqw\" (UniqueName: \"kubernetes.io/projected/bb796735-ed58-4ff1-aa33-1a8b63bc4905-kube-api-access-wpkqw\") pod \"ceilometer-0\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.064692 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.084373 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.111128 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.118331 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.121906 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.122087 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.123314 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.136468 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.246849 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7e2afa-d166-4794-9db2-db2630fd0d56-logs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.246951 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.246990 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.247025 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl5dm\" (UniqueName: \"kubernetes.io/projected/4d7e2afa-d166-4794-9db2-db2630fd0d56-kube-api-access-tl5dm\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.247119 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-config-data\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.247170 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.348833 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7e2afa-d166-4794-9db2-db2630fd0d56-logs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.348929 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.348964 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.348995 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl5dm\" (UniqueName: \"kubernetes.io/projected/4d7e2afa-d166-4794-9db2-db2630fd0d56-kube-api-access-tl5dm\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.349408 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-config-data\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.349445 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7e2afa-d166-4794-9db2-db2630fd0d56-logs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.349472 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.354701 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-config-data\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.356297 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.356652 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.356914 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.368335 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl5dm\" (UniqueName: \"kubernetes.io/projected/4d7e2afa-d166-4794-9db2-db2630fd0d56-kube-api-access-tl5dm\") pod \"nova-api-0\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.473773 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.723569 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.738579 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.860412 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-config-data\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.860615 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-ceilometer-tls-certs\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.860774 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-log-httpd\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.860821 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-scripts\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.860889 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-combined-ca-bundle\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.860948 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-run-httpd\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.861052 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpkqw\" (UniqueName: \"kubernetes.io/projected/bb796735-ed58-4ff1-aa33-1a8b63bc4905-kube-api-access-wpkqw\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.861115 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-sg-core-conf-yaml\") pod \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\" (UID: \"bb796735-ed58-4ff1-aa33-1a8b63bc4905\") " Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.861364 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.861410 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.864162 4750 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.864280 4750 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb796735-ed58-4ff1-aa33-1a8b63bc4905-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.865861 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.866733 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb796735-ed58-4ff1-aa33-1a8b63bc4905-kube-api-access-wpkqw" (OuterVolumeSpecName: "kube-api-access-wpkqw") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "kube-api-access-wpkqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.867388 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.867622 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-config-data" (OuterVolumeSpecName: "config-data") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.868141 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-scripts" (OuterVolumeSpecName: "scripts") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.868954 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb796735-ed58-4ff1-aa33-1a8b63bc4905" (UID: "bb796735-ed58-4ff1-aa33-1a8b63bc4905"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.965805 4750 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.965839 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.965849 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.965860 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpkqw\" (UniqueName: \"kubernetes.io/projected/bb796735-ed58-4ff1-aa33-1a8b63bc4905-kube-api-access-wpkqw\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.965870 4750 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.965879 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb796735-ed58-4ff1-aa33-1a8b63bc4905-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:50 crc kubenswrapper[4750]: I0309 18:50:50.970688 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.391357 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="221bc157-a4af-4e07-89fa-1bbd1dcd254c" path="/var/lib/kubelet/pods/221bc157-a4af-4e07-89fa-1bbd1dcd254c/volumes" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.392886 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c2b6a83-2323-4271-a679-85b9c533c5cc" path="/var/lib/kubelet/pods/8c2b6a83-2323-4271-a679-85b9c533c5cc/volumes" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.740597 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.740608 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d7e2afa-d166-4794-9db2-db2630fd0d56","Type":"ContainerStarted","Data":"f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea"} Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.740684 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d7e2afa-d166-4794-9db2-db2630fd0d56","Type":"ContainerStarted","Data":"dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2"} Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.740700 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d7e2afa-d166-4794-9db2-db2630fd0d56","Type":"ContainerStarted","Data":"200ab65ba8f2e2832aec42f1e97e5e1257dbb907a6d8e563ba949e4d645d85db"} Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.743291 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.743346 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.743389 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.743783 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0bf642275aba662d4b9920780da985536f4c7bcb2a54a88c1149087655cecbbd"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.743863 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://0bf642275aba662d4b9920780da985536f4c7bcb2a54a88c1149087655cecbbd" gracePeriod=600 Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.778961 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.778941443 podStartE2EDuration="1.778941443s" podCreationTimestamp="2026-03-09 18:50:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:51.773178646 +0000 UTC m=+1533.115651044" watchObservedRunningTime="2026-03-09 18:50:51.778941443 +0000 UTC m=+1533.121413831" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.829706 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.842486 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.870323 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.873442 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.877568 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.877688 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.877910 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.937346 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.991824 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60e48e88-395d-4b6d-9815-70165c7782b5-run-httpd\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.991875 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60e48e88-395d-4b6d-9815-70165c7782b5-log-httpd\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.991916 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-scripts\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.991953 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.991974 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.992007 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-config-data\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.992186 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:51 crc kubenswrapper[4750]: I0309 18:50:51.992253 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkkrx\" (UniqueName: \"kubernetes.io/projected/60e48e88-395d-4b6d-9815-70165c7782b5-kube-api-access-gkkrx\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.094853 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.094917 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-config-data\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.094956 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.094975 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkkrx\" (UniqueName: \"kubernetes.io/projected/60e48e88-395d-4b6d-9815-70165c7782b5-kube-api-access-gkkrx\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.095065 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60e48e88-395d-4b6d-9815-70165c7782b5-run-httpd\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.095090 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60e48e88-395d-4b6d-9815-70165c7782b5-log-httpd\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.095125 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-scripts\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.095158 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.096878 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60e48e88-395d-4b6d-9815-70165c7782b5-run-httpd\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.097461 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60e48e88-395d-4b6d-9815-70165c7782b5-log-httpd\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.103099 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-config-data\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.103875 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-scripts\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.103945 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.106391 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.111426 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e48e88-395d-4b6d-9815-70165c7782b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.118220 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkkrx\" (UniqueName: \"kubernetes.io/projected/60e48e88-395d-4b6d-9815-70165c7782b5-kube-api-access-gkkrx\") pod \"ceilometer-0\" (UID: \"60e48e88-395d-4b6d-9815-70165c7782b5\") " pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.201957 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.722950 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 09 18:50:52 crc kubenswrapper[4750]: W0309 18:50:52.727220 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60e48e88_395d_4b6d_9815_70165c7782b5.slice/crio-3af20da62201a37658415fd619bbab25124744d0d57b31356cdd9541232045d0 WatchSource:0}: Error finding container 3af20da62201a37658415fd619bbab25124744d0d57b31356cdd9541232045d0: Status 404 returned error can't find the container with id 3af20da62201a37658415fd619bbab25124744d0d57b31356cdd9541232045d0 Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.754089 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60e48e88-395d-4b6d-9815-70165c7782b5","Type":"ContainerStarted","Data":"3af20da62201a37658415fd619bbab25124744d0d57b31356cdd9541232045d0"} Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.757556 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="0bf642275aba662d4b9920780da985536f4c7bcb2a54a88c1149087655cecbbd" exitCode=0 Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.757693 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"0bf642275aba662d4b9920780da985536f4c7bcb2a54a88c1149087655cecbbd"} Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.757839 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1"} Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.757904 4750 scope.go:117] "RemoveContainer" containerID="aac3b01863d2a16c0edcf63137329635f54035a4884d4a4ea6dddfcacb5b31f0" Mar 09 18:50:52 crc kubenswrapper[4750]: I0309 18:50:52.989317 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:53 crc kubenswrapper[4750]: I0309 18:50:53.025742 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:53 crc kubenswrapper[4750]: I0309 18:50:53.387428 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb796735-ed58-4ff1-aa33-1a8b63bc4905" path="/var/lib/kubelet/pods/bb796735-ed58-4ff1-aa33-1a8b63bc4905/volumes" Mar 09 18:50:53 crc kubenswrapper[4750]: I0309 18:50:53.768837 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60e48e88-395d-4b6d-9815-70165c7782b5","Type":"ContainerStarted","Data":"29e931bd1e722c12b930eba631067eab99f1652780a1935202dd78047b82c0f8"} Mar 09 18:50:53 crc kubenswrapper[4750]: I0309 18:50:53.768935 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60e48e88-395d-4b6d-9815-70165c7782b5","Type":"ContainerStarted","Data":"9579065c3e773985f687a78c8eb70fa5d48fcbbae09ec1b1c52ddd685946fd28"} Mar 09 18:50:53 crc kubenswrapper[4750]: I0309 18:50:53.794331 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.026496 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-4xzzc"] Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.028114 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.030738 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.030949 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.047132 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4xzzc"] Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.141324 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.141432 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-scripts\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.141482 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkhkq\" (UniqueName: \"kubernetes.io/projected/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-kube-api-access-gkhkq\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.141567 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-config-data\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.243581 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkhkq\" (UniqueName: \"kubernetes.io/projected/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-kube-api-access-gkhkq\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.244012 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-config-data\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.244054 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.244135 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-scripts\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.248810 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-config-data\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.248874 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-scripts\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.257844 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.272224 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkhkq\" (UniqueName: \"kubernetes.io/projected/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-kube-api-access-gkhkq\") pod \"nova-cell1-cell-mapping-4xzzc\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.423180 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:50:54 crc kubenswrapper[4750]: I0309 18:50:54.800791 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60e48e88-395d-4b6d-9815-70165c7782b5","Type":"ContainerStarted","Data":"b91081f055c2aff1306457b8a6e23beb34f5894d4b812b538452f6e66bdb5242"} Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.033147 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4xzzc"] Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.164855 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.301770 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd64c67c7-xbzkw"] Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.302367 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" podUID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerName="dnsmasq-dns" containerID="cri-o://c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e" gracePeriod=10 Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.800779 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.811084 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4xzzc" event={"ID":"7c34fb79-24fb-4f5a-b7f9-09e4ae449607","Type":"ContainerStarted","Data":"6d72aa0e053da98f908c0efe2d9060f98662621d865cc3084b9e3c0451bc1878"} Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.811319 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4xzzc" event={"ID":"7c34fb79-24fb-4f5a-b7f9-09e4ae449607","Type":"ContainerStarted","Data":"112cc7be0415bb4f74f61345670bc8739a8a462683080d17caf74d7d71c167c7"} Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.823824 4750 generic.go:334] "Generic (PLEG): container finished" podID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerID="c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e" exitCode=0 Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.824046 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" event={"ID":"c5fbd87e-29ab-48a6-80a2-80a905d91607","Type":"ContainerDied","Data":"c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e"} Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.824187 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" event={"ID":"c5fbd87e-29ab-48a6-80a2-80a905d91607","Type":"ContainerDied","Data":"74f993dbf67be49ad42c316ac514dba7768a3181b93106193a2cbdf294b11d36"} Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.824215 4750 scope.go:117] "RemoveContainer" containerID="c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.828068 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd64c67c7-xbzkw" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.853537 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-4xzzc" podStartSLOduration=2.853515951 podStartE2EDuration="2.853515951s" podCreationTimestamp="2026-03-09 18:50:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:50:55.837101055 +0000 UTC m=+1537.179573453" watchObservedRunningTime="2026-03-09 18:50:55.853515951 +0000 UTC m=+1537.195988349" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.853989 4750 scope.go:117] "RemoveContainer" containerID="464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.879790 4750 scope.go:117] "RemoveContainer" containerID="c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e" Mar 09 18:50:55 crc kubenswrapper[4750]: E0309 18:50:55.880343 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e\": container with ID starting with c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e not found: ID does not exist" containerID="c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.880396 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e"} err="failed to get container status \"c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e\": rpc error: code = NotFound desc = could not find container \"c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e\": container with ID starting with c703fe22db9ee115430e7410a8789c28175d7804e2c8aea7b258c44cbc4fde0e not found: ID does not exist" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.880435 4750 scope.go:117] "RemoveContainer" containerID="464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7" Mar 09 18:50:55 crc kubenswrapper[4750]: E0309 18:50:55.881039 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7\": container with ID starting with 464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7 not found: ID does not exist" containerID="464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.881074 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7"} err="failed to get container status \"464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7\": rpc error: code = NotFound desc = could not find container \"464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7\": container with ID starting with 464e8e8fb596978b1b792229c0e821d6aab386782eafeb8dc1ac6bb9584fddc7 not found: ID does not exist" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.897921 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-config\") pod \"c5fbd87e-29ab-48a6-80a2-80a905d91607\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.898042 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-svc\") pod \"c5fbd87e-29ab-48a6-80a2-80a905d91607\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.898395 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-swift-storage-0\") pod \"c5fbd87e-29ab-48a6-80a2-80a905d91607\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.898412 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-nb\") pod \"c5fbd87e-29ab-48a6-80a2-80a905d91607\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.898456 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-sb\") pod \"c5fbd87e-29ab-48a6-80a2-80a905d91607\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.898527 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbj22\" (UniqueName: \"kubernetes.io/projected/c5fbd87e-29ab-48a6-80a2-80a905d91607-kube-api-access-mbj22\") pod \"c5fbd87e-29ab-48a6-80a2-80a905d91607\" (UID: \"c5fbd87e-29ab-48a6-80a2-80a905d91607\") " Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.911524 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5fbd87e-29ab-48a6-80a2-80a905d91607-kube-api-access-mbj22" (OuterVolumeSpecName: "kube-api-access-mbj22") pod "c5fbd87e-29ab-48a6-80a2-80a905d91607" (UID: "c5fbd87e-29ab-48a6-80a2-80a905d91607"). InnerVolumeSpecName "kube-api-access-mbj22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.965921 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c5fbd87e-29ab-48a6-80a2-80a905d91607" (UID: "c5fbd87e-29ab-48a6-80a2-80a905d91607"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.970557 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-config" (OuterVolumeSpecName: "config") pod "c5fbd87e-29ab-48a6-80a2-80a905d91607" (UID: "c5fbd87e-29ab-48a6-80a2-80a905d91607"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.976225 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c5fbd87e-29ab-48a6-80a2-80a905d91607" (UID: "c5fbd87e-29ab-48a6-80a2-80a905d91607"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.985475 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c5fbd87e-29ab-48a6-80a2-80a905d91607" (UID: "c5fbd87e-29ab-48a6-80a2-80a905d91607"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:55 crc kubenswrapper[4750]: I0309 18:50:55.989131 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c5fbd87e-29ab-48a6-80a2-80a905d91607" (UID: "c5fbd87e-29ab-48a6-80a2-80a905d91607"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.001939 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.001990 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.002001 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.002010 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.002019 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbj22\" (UniqueName: \"kubernetes.io/projected/c5fbd87e-29ab-48a6-80a2-80a905d91607-kube-api-access-mbj22\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.002031 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5fbd87e-29ab-48a6-80a2-80a905d91607-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.259245 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd64c67c7-xbzkw"] Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.274918 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd64c67c7-xbzkw"] Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.839272 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60e48e88-395d-4b6d-9815-70165c7782b5","Type":"ContainerStarted","Data":"4e51437b5230b850a380609550329a9b3e0698850369e11500de34971b8a5dec"} Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.839399 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 09 18:50:56 crc kubenswrapper[4750]: I0309 18:50:56.864028 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.537302296 podStartE2EDuration="5.864011204s" podCreationTimestamp="2026-03-09 18:50:51 +0000 UTC" firstStartedPulling="2026-03-09 18:50:52.730158902 +0000 UTC m=+1534.072631300" lastFinishedPulling="2026-03-09 18:50:56.05686781 +0000 UTC m=+1537.399340208" observedRunningTime="2026-03-09 18:50:56.862930894 +0000 UTC m=+1538.205403302" watchObservedRunningTime="2026-03-09 18:50:56.864011204 +0000 UTC m=+1538.206483602" Mar 09 18:50:57 crc kubenswrapper[4750]: I0309 18:50:57.385771 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5fbd87e-29ab-48a6-80a2-80a905d91607" path="/var/lib/kubelet/pods/c5fbd87e-29ab-48a6-80a2-80a905d91607/volumes" Mar 09 18:51:00 crc kubenswrapper[4750]: I0309 18:51:00.474446 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:51:00 crc kubenswrapper[4750]: I0309 18:51:00.475179 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:51:00 crc kubenswrapper[4750]: I0309 18:51:00.880717 4750 generic.go:334] "Generic (PLEG): container finished" podID="7c34fb79-24fb-4f5a-b7f9-09e4ae449607" containerID="6d72aa0e053da98f908c0efe2d9060f98662621d865cc3084b9e3c0451bc1878" exitCode=0 Mar 09 18:51:00 crc kubenswrapper[4750]: I0309 18:51:00.880770 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4xzzc" event={"ID":"7c34fb79-24fb-4f5a-b7f9-09e4ae449607","Type":"ContainerDied","Data":"6d72aa0e053da98f908c0efe2d9060f98662621d865cc3084b9e3c0451bc1878"} Mar 09 18:51:01 crc kubenswrapper[4750]: I0309 18:51:01.490768 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.237:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 09 18:51:01 crc kubenswrapper[4750]: I0309 18:51:01.490780 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.237:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.354237 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.444376 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-combined-ca-bundle\") pod \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.444443 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-scripts\") pod \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.444615 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-config-data\") pod \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.444771 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkhkq\" (UniqueName: \"kubernetes.io/projected/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-kube-api-access-gkhkq\") pod \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\" (UID: \"7c34fb79-24fb-4f5a-b7f9-09e4ae449607\") " Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.452801 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-kube-api-access-gkhkq" (OuterVolumeSpecName: "kube-api-access-gkhkq") pod "7c34fb79-24fb-4f5a-b7f9-09e4ae449607" (UID: "7c34fb79-24fb-4f5a-b7f9-09e4ae449607"). InnerVolumeSpecName "kube-api-access-gkhkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.454891 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-scripts" (OuterVolumeSpecName: "scripts") pod "7c34fb79-24fb-4f5a-b7f9-09e4ae449607" (UID: "7c34fb79-24fb-4f5a-b7f9-09e4ae449607"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.481811 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-config-data" (OuterVolumeSpecName: "config-data") pod "7c34fb79-24fb-4f5a-b7f9-09e4ae449607" (UID: "7c34fb79-24fb-4f5a-b7f9-09e4ae449607"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.490774 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c34fb79-24fb-4f5a-b7f9-09e4ae449607" (UID: "7c34fb79-24fb-4f5a-b7f9-09e4ae449607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.549340 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.549371 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.549381 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.549391 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkhkq\" (UniqueName: \"kubernetes.io/projected/7c34fb79-24fb-4f5a-b7f9-09e4ae449607-kube-api-access-gkhkq\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.899933 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4xzzc" event={"ID":"7c34fb79-24fb-4f5a-b7f9-09e4ae449607","Type":"ContainerDied","Data":"112cc7be0415bb4f74f61345670bc8739a8a462683080d17caf74d7d71c167c7"} Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.899979 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="112cc7be0415bb4f74f61345670bc8739a8a462683080d17caf74d7d71c167c7" Mar 09 18:51:02 crc kubenswrapper[4750]: I0309 18:51:02.900034 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4xzzc" Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.117385 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.117747 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-log" containerID="cri-o://dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2" gracePeriod=30 Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.117933 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-api" containerID="cri-o://f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea" gracePeriod=30 Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.126742 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.126984 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b2b28e7a-8d56-469f-bbd3-25524fb5a438" containerName="nova-scheduler-scheduler" containerID="cri-o://ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6" gracePeriod=30 Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.157887 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.158176 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-log" containerID="cri-o://56085c5c1f7999df142c9835426770bc49dc1dd4df13c18e365dc9d4e4662192" gracePeriod=30 Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.158322 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-metadata" containerID="cri-o://f380d429b5b30034c5ab645d5a7d07a77b9959ffdd9b3c9d697d611ce1a7672f" gracePeriod=30 Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.917004 4750 generic.go:334] "Generic (PLEG): container finished" podID="2359c431-502c-4c73-8618-d871ec51bfc9" containerID="56085c5c1f7999df142c9835426770bc49dc1dd4df13c18e365dc9d4e4662192" exitCode=143 Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.917094 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2359c431-502c-4c73-8618-d871ec51bfc9","Type":"ContainerDied","Data":"56085c5c1f7999df142c9835426770bc49dc1dd4df13c18e365dc9d4e4662192"} Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.920705 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerID="dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2" exitCode=143 Mar 09 18:51:03 crc kubenswrapper[4750]: I0309 18:51:03.920786 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d7e2afa-d166-4794-9db2-db2630fd0d56","Type":"ContainerDied","Data":"dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2"} Mar 09 18:51:04 crc kubenswrapper[4750]: I0309 18:51:04.944195 4750 generic.go:334] "Generic (PLEG): container finished" podID="2359c431-502c-4c73-8618-d871ec51bfc9" containerID="f380d429b5b30034c5ab645d5a7d07a77b9959ffdd9b3c9d697d611ce1a7672f" exitCode=0 Mar 09 18:51:04 crc kubenswrapper[4750]: I0309 18:51:04.944262 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2359c431-502c-4c73-8618-d871ec51bfc9","Type":"ContainerDied","Data":"f380d429b5b30034c5ab645d5a7d07a77b9959ffdd9b3c9d697d611ce1a7672f"} Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.307411 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.415035 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-combined-ca-bundle\") pod \"2359c431-502c-4c73-8618-d871ec51bfc9\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.415448 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-config-data\") pod \"2359c431-502c-4c73-8618-d871ec51bfc9\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.418955 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2359c431-502c-4c73-8618-d871ec51bfc9-logs\") pod \"2359c431-502c-4c73-8618-d871ec51bfc9\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.419059 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26j9x\" (UniqueName: \"kubernetes.io/projected/2359c431-502c-4c73-8618-d871ec51bfc9-kube-api-access-26j9x\") pod \"2359c431-502c-4c73-8618-d871ec51bfc9\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.419168 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-nova-metadata-tls-certs\") pod \"2359c431-502c-4c73-8618-d871ec51bfc9\" (UID: \"2359c431-502c-4c73-8618-d871ec51bfc9\") " Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.421363 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2359c431-502c-4c73-8618-d871ec51bfc9-logs" (OuterVolumeSpecName: "logs") pod "2359c431-502c-4c73-8618-d871ec51bfc9" (UID: "2359c431-502c-4c73-8618-d871ec51bfc9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.435895 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2359c431-502c-4c73-8618-d871ec51bfc9-kube-api-access-26j9x" (OuterVolumeSpecName: "kube-api-access-26j9x") pod "2359c431-502c-4c73-8618-d871ec51bfc9" (UID: "2359c431-502c-4c73-8618-d871ec51bfc9"). InnerVolumeSpecName "kube-api-access-26j9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.460830 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-config-data" (OuterVolumeSpecName: "config-data") pod "2359c431-502c-4c73-8618-d871ec51bfc9" (UID: "2359c431-502c-4c73-8618-d871ec51bfc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.467954 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2359c431-502c-4c73-8618-d871ec51bfc9" (UID: "2359c431-502c-4c73-8618-d871ec51bfc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.495778 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2359c431-502c-4c73-8618-d871ec51bfc9" (UID: "2359c431-502c-4c73-8618-d871ec51bfc9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.523002 4750 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.523024 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.523035 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2359c431-502c-4c73-8618-d871ec51bfc9-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.523044 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2359c431-502c-4c73-8618-d871ec51bfc9-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.523055 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26j9x\" (UniqueName: \"kubernetes.io/projected/2359c431-502c-4c73-8618-d871ec51bfc9-kube-api-access-26j9x\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.960895 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2359c431-502c-4c73-8618-d871ec51bfc9","Type":"ContainerDied","Data":"04f0fbda3a91280de13e5bafc5a15c80964dda3c5facbdea8e32cdb16051ea62"} Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.960955 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:51:05 crc kubenswrapper[4750]: I0309 18:51:05.960959 4750 scope.go:117] "RemoveContainer" containerID="f380d429b5b30034c5ab645d5a7d07a77b9959ffdd9b3c9d697d611ce1a7672f" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.027006 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.030185 4750 scope.go:117] "RemoveContainer" containerID="56085c5c1f7999df142c9835426770bc49dc1dd4df13c18e365dc9d4e4662192" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.040370 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.078204 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:51:06 crc kubenswrapper[4750]: E0309 18:51:06.078724 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerName="dnsmasq-dns" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.078741 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerName="dnsmasq-dns" Mar 09 18:51:06 crc kubenswrapper[4750]: E0309 18:51:06.078766 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-metadata" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.078774 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-metadata" Mar 09 18:51:06 crc kubenswrapper[4750]: E0309 18:51:06.078792 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerName="init" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.078799 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerName="init" Mar 09 18:51:06 crc kubenswrapper[4750]: E0309 18:51:06.078826 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-log" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.078833 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-log" Mar 09 18:51:06 crc kubenswrapper[4750]: E0309 18:51:06.078849 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c34fb79-24fb-4f5a-b7f9-09e4ae449607" containerName="nova-manage" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.078856 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c34fb79-24fb-4f5a-b7f9-09e4ae449607" containerName="nova-manage" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.079064 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-log" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.079085 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c34fb79-24fb-4f5a-b7f9-09e4ae449607" containerName="nova-manage" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.079103 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5fbd87e-29ab-48a6-80a2-80a905d91607" containerName="dnsmasq-dns" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.079118 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" containerName="nova-metadata-metadata" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.080441 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.093142 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.093431 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.099030 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.137590 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.137835 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f24d405-c297-4d1c-a05c-570e9d532574-logs\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.137885 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.137911 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk6bm\" (UniqueName: \"kubernetes.io/projected/5f24d405-c297-4d1c-a05c-570e9d532574-kube-api-access-bk6bm\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.137960 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-config-data\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.239764 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.239839 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f24d405-c297-4d1c-a05c-570e9d532574-logs\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.239917 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.239944 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk6bm\" (UniqueName: \"kubernetes.io/projected/5f24d405-c297-4d1c-a05c-570e9d532574-kube-api-access-bk6bm\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.240381 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f24d405-c297-4d1c-a05c-570e9d532574-logs\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.240136 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-config-data\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.247798 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.251815 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-config-data\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.252089 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f24d405-c297-4d1c-a05c-570e9d532574-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.278409 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk6bm\" (UniqueName: \"kubernetes.io/projected/5f24d405-c297-4d1c-a05c-570e9d532574-kube-api-access-bk6bm\") pod \"nova-metadata-0\" (UID: \"5f24d405-c297-4d1c-a05c-570e9d532574\") " pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.449813 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.472679 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.590319 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl5dm\" (UniqueName: \"kubernetes.io/projected/4d7e2afa-d166-4794-9db2-db2630fd0d56-kube-api-access-tl5dm\") pod \"4d7e2afa-d166-4794-9db2-db2630fd0d56\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.590703 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-public-tls-certs\") pod \"4d7e2afa-d166-4794-9db2-db2630fd0d56\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.590755 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-config-data\") pod \"4d7e2afa-d166-4794-9db2-db2630fd0d56\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.590896 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7e2afa-d166-4794-9db2-db2630fd0d56-logs\") pod \"4d7e2afa-d166-4794-9db2-db2630fd0d56\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.590948 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-combined-ca-bundle\") pod \"4d7e2afa-d166-4794-9db2-db2630fd0d56\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.591094 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-internal-tls-certs\") pod \"4d7e2afa-d166-4794-9db2-db2630fd0d56\" (UID: \"4d7e2afa-d166-4794-9db2-db2630fd0d56\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.594292 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d7e2afa-d166-4794-9db2-db2630fd0d56-logs" (OuterVolumeSpecName: "logs") pod "4d7e2afa-d166-4794-9db2-db2630fd0d56" (UID: "4d7e2afa-d166-4794-9db2-db2630fd0d56"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.596947 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7e2afa-d166-4794-9db2-db2630fd0d56-kube-api-access-tl5dm" (OuterVolumeSpecName: "kube-api-access-tl5dm") pod "4d7e2afa-d166-4794-9db2-db2630fd0d56" (UID: "4d7e2afa-d166-4794-9db2-db2630fd0d56"). InnerVolumeSpecName "kube-api-access-tl5dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.606727 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.645730 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d7e2afa-d166-4794-9db2-db2630fd0d56" (UID: "4d7e2afa-d166-4794-9db2-db2630fd0d56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.655733 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-config-data" (OuterVolumeSpecName: "config-data") pod "4d7e2afa-d166-4794-9db2-db2630fd0d56" (UID: "4d7e2afa-d166-4794-9db2-db2630fd0d56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.665587 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4d7e2afa-d166-4794-9db2-db2630fd0d56" (UID: "4d7e2afa-d166-4794-9db2-db2630fd0d56"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.668127 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4d7e2afa-d166-4794-9db2-db2630fd0d56" (UID: "4d7e2afa-d166-4794-9db2-db2630fd0d56"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.693184 4750 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.693228 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl5dm\" (UniqueName: \"kubernetes.io/projected/4d7e2afa-d166-4794-9db2-db2630fd0d56-kube-api-access-tl5dm\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.693244 4750 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.693253 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.693263 4750 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7e2afa-d166-4794-9db2-db2630fd0d56-logs\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.693271 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7e2afa-d166-4794-9db2-db2630fd0d56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.794405 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7klpv\" (UniqueName: \"kubernetes.io/projected/b2b28e7a-8d56-469f-bbd3-25524fb5a438-kube-api-access-7klpv\") pod \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.794899 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-combined-ca-bundle\") pod \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.795125 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-config-data\") pod \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\" (UID: \"b2b28e7a-8d56-469f-bbd3-25524fb5a438\") " Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.797857 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b28e7a-8d56-469f-bbd3-25524fb5a438-kube-api-access-7klpv" (OuterVolumeSpecName: "kube-api-access-7klpv") pod "b2b28e7a-8d56-469f-bbd3-25524fb5a438" (UID: "b2b28e7a-8d56-469f-bbd3-25524fb5a438"). InnerVolumeSpecName "kube-api-access-7klpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.826233 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-config-data" (OuterVolumeSpecName: "config-data") pod "b2b28e7a-8d56-469f-bbd3-25524fb5a438" (UID: "b2b28e7a-8d56-469f-bbd3-25524fb5a438"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.826600 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2b28e7a-8d56-469f-bbd3-25524fb5a438" (UID: "b2b28e7a-8d56-469f-bbd3-25524fb5a438"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.898063 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.898097 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2b28e7a-8d56-469f-bbd3-25524fb5a438-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.898107 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7klpv\" (UniqueName: \"kubernetes.io/projected/b2b28e7a-8d56-469f-bbd3-25524fb5a438-kube-api-access-7klpv\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.967271 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.983339 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerID="f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea" exitCode=0 Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.983442 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.983441 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d7e2afa-d166-4794-9db2-db2630fd0d56","Type":"ContainerDied","Data":"f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea"} Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.983524 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d7e2afa-d166-4794-9db2-db2630fd0d56","Type":"ContainerDied","Data":"200ab65ba8f2e2832aec42f1e97e5e1257dbb907a6d8e563ba949e4d645d85db"} Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.983553 4750 scope.go:117] "RemoveContainer" containerID="f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.985712 4750 generic.go:334] "Generic (PLEG): container finished" podID="b2b28e7a-8d56-469f-bbd3-25524fb5a438" containerID="ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6" exitCode=0 Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.985785 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2b28e7a-8d56-469f-bbd3-25524fb5a438","Type":"ContainerDied","Data":"ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6"} Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.985806 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2b28e7a-8d56-469f-bbd3-25524fb5a438","Type":"ContainerDied","Data":"5470a992ec245a9c25a8a2b2b46d69dc68d92b9f5ed763d78f73ea3de4dee81c"} Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.985868 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:51:06 crc kubenswrapper[4750]: I0309 18:51:06.996529 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f24d405-c297-4d1c-a05c-570e9d532574","Type":"ContainerStarted","Data":"45367380ce1d029b62591508d83180d55e21a19e5a7294e63a8f83c218c14385"} Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.034448 4750 scope.go:117] "RemoveContainer" containerID="dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.040410 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.079380 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.100707 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.100794 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.111564 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: E0309 18:51:07.112102 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-api" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.112124 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-api" Mar 09 18:51:07 crc kubenswrapper[4750]: E0309 18:51:07.112142 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-log" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.112149 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-log" Mar 09 18:51:07 crc kubenswrapper[4750]: E0309 18:51:07.112160 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b28e7a-8d56-469f-bbd3-25524fb5a438" containerName="nova-scheduler-scheduler" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.112168 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b28e7a-8d56-469f-bbd3-25524fb5a438" containerName="nova-scheduler-scheduler" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.112384 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-log" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.112421 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b28e7a-8d56-469f-bbd3-25524fb5a438" containerName="nova-scheduler-scheduler" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.112434 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" containerName="nova-api-api" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.113622 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.115708 4750 scope.go:117] "RemoveContainer" containerID="f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea" Mar 09 18:51:07 crc kubenswrapper[4750]: E0309 18:51:07.116158 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea\": container with ID starting with f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea not found: ID does not exist" containerID="f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.116315 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea"} err="failed to get container status \"f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea\": rpc error: code = NotFound desc = could not find container \"f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea\": container with ID starting with f825282f4033605097e8041269b9a28fd2a30db688f0029c37208cab1efbacea not found: ID does not exist" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.116480 4750 scope.go:117] "RemoveContainer" containerID="dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2" Mar 09 18:51:07 crc kubenswrapper[4750]: E0309 18:51:07.116782 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2\": container with ID starting with dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2 not found: ID does not exist" containerID="dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.116817 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2"} err="failed to get container status \"dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2\": rpc error: code = NotFound desc = could not find container \"dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2\": container with ID starting with dd22003e43c4e31928a3829634acc44f7e1173062fab229a0879f8bf10abd4f2 not found: ID does not exist" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.116835 4750 scope.go:117] "RemoveContainer" containerID="ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.116888 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.117442 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.117768 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.122984 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.124376 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.131096 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.134676 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.148689 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.153224 4750 scope.go:117] "RemoveContainer" containerID="ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6" Mar 09 18:51:07 crc kubenswrapper[4750]: E0309 18:51:07.154155 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6\": container with ID starting with ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6 not found: ID does not exist" containerID="ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.154189 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6"} err="failed to get container status \"ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6\": rpc error: code = NotFound desc = could not find container \"ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6\": container with ID starting with ce6e687a3d5daf056b325574043cd8a19c0d7afca3378d9f2a24da4d1b2c9ce6 not found: ID does not exist" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.203660 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.203741 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50d1b356-0b06-47eb-9fdb-6674038e109d-logs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.203820 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdcdm\" (UniqueName: \"kubernetes.io/projected/50d1b356-0b06-47eb-9fdb-6674038e109d-kube-api-access-gdcdm\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.203884 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.203913 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-config-data\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.204218 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-public-tls-certs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.305993 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306069 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cb5c89-cbc8-469d-afe9-45100b0867e0-config-data\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306108 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50d1b356-0b06-47eb-9fdb-6674038e109d-logs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306145 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdcdm\" (UniqueName: \"kubernetes.io/projected/50d1b356-0b06-47eb-9fdb-6674038e109d-kube-api-access-gdcdm\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306277 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306361 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-config-data\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306405 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cb5c89-cbc8-469d-afe9-45100b0867e0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306523 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlhht\" (UniqueName: \"kubernetes.io/projected/d4cb5c89-cbc8-469d-afe9-45100b0867e0-kube-api-access-wlhht\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306658 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-public-tls-certs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.306670 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50d1b356-0b06-47eb-9fdb-6674038e109d-logs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.309800 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.310605 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-config-data\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.311358 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-public-tls-certs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.311247 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50d1b356-0b06-47eb-9fdb-6674038e109d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.322512 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdcdm\" (UniqueName: \"kubernetes.io/projected/50d1b356-0b06-47eb-9fdb-6674038e109d-kube-api-access-gdcdm\") pod \"nova-api-0\" (UID: \"50d1b356-0b06-47eb-9fdb-6674038e109d\") " pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.386375 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2359c431-502c-4c73-8618-d871ec51bfc9" path="/var/lib/kubelet/pods/2359c431-502c-4c73-8618-d871ec51bfc9/volumes" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.387081 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7e2afa-d166-4794-9db2-db2630fd0d56" path="/var/lib/kubelet/pods/4d7e2afa-d166-4794-9db2-db2630fd0d56/volumes" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.387650 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b28e7a-8d56-469f-bbd3-25524fb5a438" path="/var/lib/kubelet/pods/b2b28e7a-8d56-469f-bbd3-25524fb5a438/volumes" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.409165 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cb5c89-cbc8-469d-afe9-45100b0867e0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.409250 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlhht\" (UniqueName: \"kubernetes.io/projected/d4cb5c89-cbc8-469d-afe9-45100b0867e0-kube-api-access-wlhht\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.409417 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cb5c89-cbc8-469d-afe9-45100b0867e0-config-data\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.412874 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cb5c89-cbc8-469d-afe9-45100b0867e0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.413258 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cb5c89-cbc8-469d-afe9-45100b0867e0-config-data\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.429552 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlhht\" (UniqueName: \"kubernetes.io/projected/d4cb5c89-cbc8-469d-afe9-45100b0867e0-kube-api-access-wlhht\") pod \"nova-scheduler-0\" (UID: \"d4cb5c89-cbc8-469d-afe9-45100b0867e0\") " pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.476172 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.489714 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 09 18:51:07 crc kubenswrapper[4750]: W0309 18:51:07.969206 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4cb5c89_cbc8_469d_afe9_45100b0867e0.slice/crio-6b5f27ec6918e9242ba15e96d9c84c3b87d44c4924be617eeeec55c69b81deac WatchSource:0}: Error finding container 6b5f27ec6918e9242ba15e96d9c84c3b87d44c4924be617eeeec55c69b81deac: Status 404 returned error can't find the container with id 6b5f27ec6918e9242ba15e96d9c84c3b87d44c4924be617eeeec55c69b81deac Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.971202 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 09 18:51:07 crc kubenswrapper[4750]: W0309 18:51:07.992806 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50d1b356_0b06_47eb_9fdb_6674038e109d.slice/crio-4f0a810c8899e7e225859daf41d381c0eb543b93dd0186794e1682000399e0e0 WatchSource:0}: Error finding container 4f0a810c8899e7e225859daf41d381c0eb543b93dd0186794e1682000399e0e0: Status 404 returned error can't find the container with id 4f0a810c8899e7e225859daf41d381c0eb543b93dd0186794e1682000399e0e0 Mar 09 18:51:07 crc kubenswrapper[4750]: I0309 18:51:07.994886 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 09 18:51:08 crc kubenswrapper[4750]: I0309 18:51:08.017978 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4cb5c89-cbc8-469d-afe9-45100b0867e0","Type":"ContainerStarted","Data":"6b5f27ec6918e9242ba15e96d9c84c3b87d44c4924be617eeeec55c69b81deac"} Mar 09 18:51:08 crc kubenswrapper[4750]: I0309 18:51:08.020428 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f24d405-c297-4d1c-a05c-570e9d532574","Type":"ContainerStarted","Data":"075a7c2341c6e2b0334e38827f5df797d418ce6394e22ece99a290d9d9bb3630"} Mar 09 18:51:08 crc kubenswrapper[4750]: I0309 18:51:08.020460 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f24d405-c297-4d1c-a05c-570e9d532574","Type":"ContainerStarted","Data":"d0b6b5110184513c35cec13db65f6a1633f5e4fda0bafa84382f1bd0dc56bb8b"} Mar 09 18:51:08 crc kubenswrapper[4750]: I0309 18:51:08.025321 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50d1b356-0b06-47eb-9fdb-6674038e109d","Type":"ContainerStarted","Data":"4f0a810c8899e7e225859daf41d381c0eb543b93dd0186794e1682000399e0e0"} Mar 09 18:51:08 crc kubenswrapper[4750]: I0309 18:51:08.048902 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.048879964 podStartE2EDuration="2.048879964s" podCreationTimestamp="2026-03-09 18:51:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:51:08.038577753 +0000 UTC m=+1549.381050151" watchObservedRunningTime="2026-03-09 18:51:08.048879964 +0000 UTC m=+1549.391352362" Mar 09 18:51:09 crc kubenswrapper[4750]: I0309 18:51:09.036578 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50d1b356-0b06-47eb-9fdb-6674038e109d","Type":"ContainerStarted","Data":"c17c59e707b1d19b0981001a20489a0e4c457cec2ddf338560783d1dad26edd8"} Mar 09 18:51:09 crc kubenswrapper[4750]: I0309 18:51:09.036877 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50d1b356-0b06-47eb-9fdb-6674038e109d","Type":"ContainerStarted","Data":"c42a7e2e095478c62ed3eb1cd4708d2ad4231a759aaf5a3777c1ea2e8aaf3216"} Mar 09 18:51:09 crc kubenswrapper[4750]: I0309 18:51:09.039106 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d4cb5c89-cbc8-469d-afe9-45100b0867e0","Type":"ContainerStarted","Data":"1e66d563719b94930ed0dfb451ef4cd31c53a8072168891e8d5091d248da1830"} Mar 09 18:51:09 crc kubenswrapper[4750]: I0309 18:51:09.073871 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.073851641 podStartE2EDuration="2.073851641s" podCreationTimestamp="2026-03-09 18:51:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:51:09.064799324 +0000 UTC m=+1550.407271722" watchObservedRunningTime="2026-03-09 18:51:09.073851641 +0000 UTC m=+1550.416324039" Mar 09 18:51:09 crc kubenswrapper[4750]: I0309 18:51:09.096285 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.096264072 podStartE2EDuration="2.096264072s" podCreationTimestamp="2026-03-09 18:51:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:51:09.086696651 +0000 UTC m=+1550.429169049" watchObservedRunningTime="2026-03-09 18:51:09.096264072 +0000 UTC m=+1550.438736470" Mar 09 18:51:11 crc kubenswrapper[4750]: I0309 18:51:11.451515 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 09 18:51:11 crc kubenswrapper[4750]: I0309 18:51:11.452128 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 09 18:51:12 crc kubenswrapper[4750]: I0309 18:51:12.490077 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 09 18:51:16 crc kubenswrapper[4750]: I0309 18:51:16.451785 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 09 18:51:16 crc kubenswrapper[4750]: I0309 18:51:16.451886 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 09 18:51:17 crc kubenswrapper[4750]: I0309 18:51:17.464888 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5f24d405-c297-4d1c-a05c-570e9d532574" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.240:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 09 18:51:17 crc kubenswrapper[4750]: I0309 18:51:17.464963 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5f24d405-c297-4d1c-a05c-570e9d532574" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.240:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 09 18:51:17 crc kubenswrapper[4750]: I0309 18:51:17.476427 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:51:17 crc kubenswrapper[4750]: I0309 18:51:17.476466 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 09 18:51:17 crc kubenswrapper[4750]: I0309 18:51:17.490050 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 09 18:51:17 crc kubenswrapper[4750]: I0309 18:51:17.532577 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 09 18:51:18 crc kubenswrapper[4750]: I0309 18:51:18.194196 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 09 18:51:18 crc kubenswrapper[4750]: I0309 18:51:18.494903 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="50d1b356-0b06-47eb-9fdb-6674038e109d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.241:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 09 18:51:18 crc kubenswrapper[4750]: I0309 18:51:18.494897 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="50d1b356-0b06-47eb-9fdb-6674038e109d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.241:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 09 18:51:22 crc kubenswrapper[4750]: I0309 18:51:22.217551 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 09 18:51:26 crc kubenswrapper[4750]: I0309 18:51:26.459598 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 09 18:51:26 crc kubenswrapper[4750]: I0309 18:51:26.461475 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 09 18:51:26 crc kubenswrapper[4750]: I0309 18:51:26.469018 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 09 18:51:27 crc kubenswrapper[4750]: I0309 18:51:27.314090 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 09 18:51:27 crc kubenswrapper[4750]: I0309 18:51:27.494999 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 09 18:51:27 crc kubenswrapper[4750]: I0309 18:51:27.495458 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 09 18:51:27 crc kubenswrapper[4750]: I0309 18:51:27.495949 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 09 18:51:27 crc kubenswrapper[4750]: I0309 18:51:27.507304 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 09 18:51:28 crc kubenswrapper[4750]: I0309 18:51:28.308943 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 09 18:51:28 crc kubenswrapper[4750]: I0309 18:51:28.321436 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 09 18:51:37 crc kubenswrapper[4750]: I0309 18:51:37.325611 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:51:38 crc kubenswrapper[4750]: I0309 18:51:38.318934 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:51:41 crc kubenswrapper[4750]: I0309 18:51:41.024202 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="378d9b58-f830-4d61-b408-26668a301507" containerName="rabbitmq" containerID="cri-o://3ffaf36e1a2a216af73d4039972485467b465dcd9d2950724768a1894bcb9519" gracePeriod=604797 Mar 09 18:51:41 crc kubenswrapper[4750]: I0309 18:51:41.680033 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" containerName="rabbitmq" containerID="cri-o://ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6" gracePeriod=604797 Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.494021 4750 generic.go:334] "Generic (PLEG): container finished" podID="378d9b58-f830-4d61-b408-26668a301507" containerID="3ffaf36e1a2a216af73d4039972485467b465dcd9d2950724768a1894bcb9519" exitCode=0 Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.494163 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"378d9b58-f830-4d61-b408-26668a301507","Type":"ContainerDied","Data":"3ffaf36e1a2a216af73d4039972485467b465dcd9d2950724768a1894bcb9519"} Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.684765 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.810556 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/378d9b58-f830-4d61-b408-26668a301507-erlang-cookie-secret\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.810653 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-tls\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.810703 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-server-conf\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.810751 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-config-data\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.810802 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-erlang-cookie\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.810833 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njwwt\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-kube-api-access-njwwt\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.810947 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/378d9b58-f830-4d61-b408-26668a301507-pod-info\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.811112 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-plugins-conf\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.811200 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-confd\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.811257 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.811306 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-plugins\") pod \"378d9b58-f830-4d61-b408-26668a301507\" (UID: \"378d9b58-f830-4d61-b408-26668a301507\") " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.812515 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.814685 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.817182 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.838439 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/378d9b58-f830-4d61-b408-26668a301507-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.838456 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/378d9b58-f830-4d61-b408-26668a301507-pod-info" (OuterVolumeSpecName: "pod-info") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.838527 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.838570 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-kube-api-access-njwwt" (OuterVolumeSpecName: "kube-api-access-njwwt") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "kube-api-access-njwwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.839098 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.872413 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-config-data" (OuterVolumeSpecName: "config-data") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.881438 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-server-conf" (OuterVolumeSpecName: "server-conf") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.913557 4750 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.913611 4750 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914322 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914423 4750 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/378d9b58-f830-4d61-b408-26668a301507-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914479 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914532 4750 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-server-conf\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914583 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/378d9b58-f830-4d61-b408-26668a301507-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914653 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914741 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njwwt\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-kube-api-access-njwwt\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.914812 4750 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/378d9b58-f830-4d61-b408-26668a301507-pod-info\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.972811 4750 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 09 18:51:42 crc kubenswrapper[4750]: I0309 18:51:42.979789 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "378d9b58-f830-4d61-b408-26668a301507" (UID: "378d9b58-f830-4d61-b408-26668a301507"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.018070 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/378d9b58-f830-4d61-b408-26668a301507-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.018098 4750 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.212774 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.322651 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-confd\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.322759 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.322802 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-erlang-cookie\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.322852 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3236a6fb-c288-441c-96e5-7941f818b0af-pod-info\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.322903 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qkgb\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-kube-api-access-6qkgb\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.322956 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-server-conf\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.323054 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-plugins-conf\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.323115 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-config-data\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.323188 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3236a6fb-c288-441c-96e5-7941f818b0af-erlang-cookie-secret\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.323219 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-plugins\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.323284 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-tls\") pod \"3236a6fb-c288-441c-96e5-7941f818b0af\" (UID: \"3236a6fb-c288-441c-96e5-7941f818b0af\") " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.327261 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.328332 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.329736 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.330346 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.331665 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.333041 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-kube-api-access-6qkgb" (OuterVolumeSpecName: "kube-api-access-6qkgb") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "kube-api-access-6qkgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.341364 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3236a6fb-c288-441c-96e5-7941f818b0af-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.350401 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3236a6fb-c288-441c-96e5-7941f818b0af-pod-info" (OuterVolumeSpecName: "pod-info") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.403302 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-config-data" (OuterVolumeSpecName: "config-data") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.415960 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-server-conf" (OuterVolumeSpecName: "server-conf") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426081 4750 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3236a6fb-c288-441c-96e5-7941f818b0af-pod-info\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426113 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qkgb\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-kube-api-access-6qkgb\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426123 4750 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-server-conf\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426131 4750 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426141 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3236a6fb-c288-441c-96e5-7941f818b0af-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426151 4750 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3236a6fb-c288-441c-96e5-7941f818b0af-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426158 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426166 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426193 4750 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.426203 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.479872 4750 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.502917 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3236a6fb-c288-441c-96e5-7941f818b0af" (UID: "3236a6fb-c288-441c-96e5-7941f818b0af"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.509020 4750 generic.go:334] "Generic (PLEG): container finished" podID="3236a6fb-c288-441c-96e5-7941f818b0af" containerID="ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6" exitCode=0 Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.509138 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.512407 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.530114 4750 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3236a6fb-c288-441c-96e5-7941f818b0af-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.530143 4750 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.539802 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3236a6fb-c288-441c-96e5-7941f818b0af","Type":"ContainerDied","Data":"ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6"} Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.539955 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3236a6fb-c288-441c-96e5-7941f818b0af","Type":"ContainerDied","Data":"b6ba2eed96cadb188db02a1152302b01183198beb4c60c2b24d08166f70fd40a"} Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.540034 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"378d9b58-f830-4d61-b408-26668a301507","Type":"ContainerDied","Data":"7b7a5f2b990282f78ad5069ff7216076ea76c26aca9c00a15b6a5a84c265577b"} Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.539996 4750 scope.go:117] "RemoveContainer" containerID="ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.580382 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.591715 4750 scope.go:117] "RemoveContainer" containerID="4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.602649 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.638208 4750 scope.go:117] "RemoveContainer" containerID="ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6" Mar 09 18:51:43 crc kubenswrapper[4750]: E0309 18:51:43.639272 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6\": container with ID starting with ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6 not found: ID does not exist" containerID="ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.639322 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6"} err="failed to get container status \"ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6\": rpc error: code = NotFound desc = could not find container \"ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6\": container with ID starting with ca66d5abdb26a50619e9dd1674e99a96dc102704dce7cc0ba4c467919e9ac1b6 not found: ID does not exist" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.639351 4750 scope.go:117] "RemoveContainer" containerID="4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143" Mar 09 18:51:43 crc kubenswrapper[4750]: E0309 18:51:43.640943 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143\": container with ID starting with 4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143 not found: ID does not exist" containerID="4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.640998 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143"} err="failed to get container status \"4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143\": rpc error: code = NotFound desc = could not find container \"4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143\": container with ID starting with 4f1f71da5c5f3555601d302372f08ca88a2a539cdc8d1da99672d5001309c143 not found: ID does not exist" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.641034 4750 scope.go:117] "RemoveContainer" containerID="3ffaf36e1a2a216af73d4039972485467b465dcd9d2950724768a1894bcb9519" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.647021 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.658750 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.692679 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: E0309 18:51:43.693178 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" containerName="setup-container" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.693193 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" containerName="setup-container" Mar 09 18:51:43 crc kubenswrapper[4750]: E0309 18:51:43.693210 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" containerName="rabbitmq" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.693216 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" containerName="rabbitmq" Mar 09 18:51:43 crc kubenswrapper[4750]: E0309 18:51:43.693231 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="378d9b58-f830-4d61-b408-26668a301507" containerName="rabbitmq" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.693237 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="378d9b58-f830-4d61-b408-26668a301507" containerName="rabbitmq" Mar 09 18:51:43 crc kubenswrapper[4750]: E0309 18:51:43.693254 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="378d9b58-f830-4d61-b408-26668a301507" containerName="setup-container" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.693261 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="378d9b58-f830-4d61-b408-26668a301507" containerName="setup-container" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.693446 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" containerName="rabbitmq" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.693464 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="378d9b58-f830-4d61-b408-26668a301507" containerName="rabbitmq" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.694568 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.697031 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.697188 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.697367 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.697369 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.697533 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.697923 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.699156 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ck6bg" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.704594 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.717372 4750 scope.go:117] "RemoveContainer" containerID="bbc618a451438e977e10a76341e34bb8d2c1f770d1a8c8ffc94ff70b158dfb10" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.721672 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.724021 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.726412 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5prvj" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.726809 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.726883 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.726937 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.727075 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.727083 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.727113 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.734158 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.838244 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.838547 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4103fad-b50b-450d-99c0-38a505bd800a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.838711 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.838828 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.838903 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4103fad-b50b-450d-99c0-38a505bd800a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839015 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr56h\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-kube-api-access-dr56h\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839127 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/14f2384b-45f7-45ca-ac90-429fc2f69bbb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839209 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839308 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839394 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839553 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839653 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839699 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-config-data\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839769 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839823 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839839 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzfj5\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-kube-api-access-kzfj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839919 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.839941 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.840002 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.840023 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.840078 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/14f2384b-45f7-45ca-ac90-429fc2f69bbb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.840093 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.941858 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr56h\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-kube-api-access-dr56h\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.941906 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/14f2384b-45f7-45ca-ac90-429fc2f69bbb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.941931 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.941965 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.941987 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942013 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942035 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942058 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-config-data\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942084 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942107 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942122 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzfj5\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-kube-api-access-kzfj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942155 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942206 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942225 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942250 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/14f2384b-45f7-45ca-ac90-429fc2f69bbb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942267 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942313 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4103fad-b50b-450d-99c0-38a505bd800a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942352 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942369 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942391 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4103fad-b50b-450d-99c0-38a505bd800a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.942907 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.943490 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.943664 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.943870 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-config-data\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.944153 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.944751 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.945035 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.945385 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.945713 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.946018 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.946070 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/14f2384b-45f7-45ca-ac90-429fc2f69bbb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.946530 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4103fad-b50b-450d-99c0-38a505bd800a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.947820 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4103fad-b50b-450d-99c0-38a505bd800a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.950362 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.950718 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4103fad-b50b-450d-99c0-38a505bd800a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.952995 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/14f2384b-45f7-45ca-ac90-429fc2f69bbb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.953181 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/14f2384b-45f7-45ca-ac90-429fc2f69bbb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.954516 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.954506 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.954524 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.960357 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr56h\" (UniqueName: \"kubernetes.io/projected/14f2384b-45f7-45ca-ac90-429fc2f69bbb-kube-api-access-dr56h\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:43 crc kubenswrapper[4750]: I0309 18:51:43.963703 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzfj5\" (UniqueName: \"kubernetes.io/projected/b4103fad-b50b-450d-99c0-38a505bd800a-kube-api-access-kzfj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:44 crc kubenswrapper[4750]: I0309 18:51:44.022703 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4103fad-b50b-450d-99c0-38a505bd800a\") " pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:44 crc kubenswrapper[4750]: I0309 18:51:44.023687 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"14f2384b-45f7-45ca-ac90-429fc2f69bbb\") " pod="openstack/rabbitmq-server-0" Mar 09 18:51:44 crc kubenswrapper[4750]: I0309 18:51:44.051207 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 09 18:51:44 crc kubenswrapper[4750]: I0309 18:51:44.316658 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:51:44 crc kubenswrapper[4750]: I0309 18:51:44.507918 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 09 18:51:44 crc kubenswrapper[4750]: I0309 18:51:44.783421 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 09 18:51:44 crc kubenswrapper[4750]: W0309 18:51:44.784772 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4103fad_b50b_450d_99c0_38a505bd800a.slice/crio-fa68e92e8dbf3374a43847583800cdeb781e19ab024dbb0a61a754db429349b9 WatchSource:0}: Error finding container fa68e92e8dbf3374a43847583800cdeb781e19ab024dbb0a61a754db429349b9: Status 404 returned error can't find the container with id fa68e92e8dbf3374a43847583800cdeb781e19ab024dbb0a61a754db429349b9 Mar 09 18:51:45 crc kubenswrapper[4750]: I0309 18:51:45.390371 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3236a6fb-c288-441c-96e5-7941f818b0af" path="/var/lib/kubelet/pods/3236a6fb-c288-441c-96e5-7941f818b0af/volumes" Mar 09 18:51:45 crc kubenswrapper[4750]: I0309 18:51:45.391619 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="378d9b58-f830-4d61-b408-26668a301507" path="/var/lib/kubelet/pods/378d9b58-f830-4d61-b408-26668a301507/volumes" Mar 09 18:51:45 crc kubenswrapper[4750]: I0309 18:51:45.555799 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4103fad-b50b-450d-99c0-38a505bd800a","Type":"ContainerStarted","Data":"fa68e92e8dbf3374a43847583800cdeb781e19ab024dbb0a61a754db429349b9"} Mar 09 18:51:45 crc kubenswrapper[4750]: I0309 18:51:45.557145 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"14f2384b-45f7-45ca-ac90-429fc2f69bbb","Type":"ContainerStarted","Data":"c8c5956d7e6a7942072cefc45c2972f57bb3884ff7476ba4f705f995a8bdd709"} Mar 09 18:51:46 crc kubenswrapper[4750]: I0309 18:51:46.568283 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"14f2384b-45f7-45ca-ac90-429fc2f69bbb","Type":"ContainerStarted","Data":"99bef4fd42c532835b9aa16a998190da9547d9c841d5438d0c51e7a1a7992b8c"} Mar 09 18:51:47 crc kubenswrapper[4750]: I0309 18:51:47.593504 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4103fad-b50b-450d-99c0-38a505bd800a","Type":"ContainerStarted","Data":"1358bbd2528c010830a89b315b8dd22a7801a0040a42c2a6fbca19f26fee7d45"} Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.597969 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-859dbb75b9-2sdgv"] Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.601014 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.607018 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.619723 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-859dbb75b9-2sdgv"] Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.732645 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwqss\" (UniqueName: \"kubernetes.io/projected/e3f3d445-e629-4668-a6f7-d8b44406e823-kube-api-access-nwqss\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.732837 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-config\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.732871 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-sb\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.732903 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-svc\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.732922 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-swift-storage-0\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.733018 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-openstack-edpm-ipam\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.733042 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-nb\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.823853 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859dbb75b9-2sdgv"] Mar 09 18:51:51 crc kubenswrapper[4750]: E0309 18:51:51.824838 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-nwqss openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" podUID="e3f3d445-e629-4668-a6f7-d8b44406e823" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.835293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-svc\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.835354 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-swift-storage-0\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.835461 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-openstack-edpm-ipam\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.835485 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-nb\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.835555 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwqss\" (UniqueName: \"kubernetes.io/projected/e3f3d445-e629-4668-a6f7-d8b44406e823-kube-api-access-nwqss\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.835738 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-config\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.835771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-sb\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.836554 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-swift-storage-0\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.836566 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-openstack-edpm-ipam\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.837267 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-nb\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.837465 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-config\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.838474 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-sb\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.838804 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-svc\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.861574 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwqss\" (UniqueName: \"kubernetes.io/projected/e3f3d445-e629-4668-a6f7-d8b44406e823-kube-api-access-nwqss\") pod \"dnsmasq-dns-859dbb75b9-2sdgv\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.876658 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56d7fdf8b5-4kbkb"] Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.878861 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:51 crc kubenswrapper[4750]: I0309 18:51:51.897601 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56d7fdf8b5-4kbkb"] Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.039669 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-nb\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.039992 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7bns\" (UniqueName: \"kubernetes.io/projected/ed71b380-e6d0-4f58-be59-c01a8de14ec5-kube-api-access-n7bns\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.040548 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-svc\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.040681 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-sb\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.040795 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-swift-storage-0\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.040897 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-config\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.041002 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-openstack-edpm-ipam\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.144133 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-openstack-edpm-ipam\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.144323 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-nb\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.144400 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7bns\" (UniqueName: \"kubernetes.io/projected/ed71b380-e6d0-4f58-be59-c01a8de14ec5-kube-api-access-n7bns\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.144487 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-svc\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.144578 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-sb\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.145135 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-openstack-edpm-ipam\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.144791 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-swift-storage-0\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.145477 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-config\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.145549 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-sb\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.145595 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-nb\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.145766 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-svc\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.145792 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-swift-storage-0\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.146156 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-config\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.162217 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7bns\" (UniqueName: \"kubernetes.io/projected/ed71b380-e6d0-4f58-be59-c01a8de14ec5-kube-api-access-n7bns\") pod \"dnsmasq-dns-56d7fdf8b5-4kbkb\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.231690 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.647234 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.662150 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.705327 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56d7fdf8b5-4kbkb"] Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861112 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-openstack-edpm-ipam\") pod \"e3f3d445-e629-4668-a6f7-d8b44406e823\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861162 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-svc\") pod \"e3f3d445-e629-4668-a6f7-d8b44406e823\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861193 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-sb\") pod \"e3f3d445-e629-4668-a6f7-d8b44406e823\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861266 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-swift-storage-0\") pod \"e3f3d445-e629-4668-a6f7-d8b44406e823\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861353 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-config\") pod \"e3f3d445-e629-4668-a6f7-d8b44406e823\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861378 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwqss\" (UniqueName: \"kubernetes.io/projected/e3f3d445-e629-4668-a6f7-d8b44406e823-kube-api-access-nwqss\") pod \"e3f3d445-e629-4668-a6f7-d8b44406e823\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861404 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-nb\") pod \"e3f3d445-e629-4668-a6f7-d8b44406e823\" (UID: \"e3f3d445-e629-4668-a6f7-d8b44406e823\") " Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861713 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e3f3d445-e629-4668-a6f7-d8b44406e823" (UID: "e3f3d445-e629-4668-a6f7-d8b44406e823"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.861726 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e3f3d445-e629-4668-a6f7-d8b44406e823" (UID: "e3f3d445-e629-4668-a6f7-d8b44406e823"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.862067 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-config" (OuterVolumeSpecName: "config") pod "e3f3d445-e629-4668-a6f7-d8b44406e823" (UID: "e3f3d445-e629-4668-a6f7-d8b44406e823"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.862545 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e3f3d445-e629-4668-a6f7-d8b44406e823" (UID: "e3f3d445-e629-4668-a6f7-d8b44406e823"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.862697 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "e3f3d445-e629-4668-a6f7-d8b44406e823" (UID: "e3f3d445-e629-4668-a6f7-d8b44406e823"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.862745 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e3f3d445-e629-4668-a6f7-d8b44406e823" (UID: "e3f3d445-e629-4668-a6f7-d8b44406e823"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.866482 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3f3d445-e629-4668-a6f7-d8b44406e823-kube-api-access-nwqss" (OuterVolumeSpecName: "kube-api-access-nwqss") pod "e3f3d445-e629-4668-a6f7-d8b44406e823" (UID: "e3f3d445-e629-4668-a6f7-d8b44406e823"). InnerVolumeSpecName "kube-api-access-nwqss". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.962684 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.962925 4750 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.963017 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.963070 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.963123 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.963172 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3d445-e629-4668-a6f7-d8b44406e823-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:52 crc kubenswrapper[4750]: I0309 18:51:52.963222 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwqss\" (UniqueName: \"kubernetes.io/projected/e3f3d445-e629-4668-a6f7-d8b44406e823-kube-api-access-nwqss\") on node \"crc\" DevicePath \"\"" Mar 09 18:51:53 crc kubenswrapper[4750]: I0309 18:51:53.661725 4750 generic.go:334] "Generic (PLEG): container finished" podID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerID="4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3" exitCode=0 Mar 09 18:51:53 crc kubenswrapper[4750]: I0309 18:51:53.663139 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859dbb75b9-2sdgv" Mar 09 18:51:53 crc kubenswrapper[4750]: I0309 18:51:53.663366 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" event={"ID":"ed71b380-e6d0-4f58-be59-c01a8de14ec5","Type":"ContainerDied","Data":"4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3"} Mar 09 18:51:53 crc kubenswrapper[4750]: I0309 18:51:53.663415 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" event={"ID":"ed71b380-e6d0-4f58-be59-c01a8de14ec5","Type":"ContainerStarted","Data":"cdd52e366ed5c414c71b73a7a44f1dcfa8f5ea59cfd5b0106f2c879b027e7f42"} Mar 09 18:51:53 crc kubenswrapper[4750]: I0309 18:51:53.903308 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859dbb75b9-2sdgv"] Mar 09 18:51:53 crc kubenswrapper[4750]: I0309 18:51:53.920215 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-859dbb75b9-2sdgv"] Mar 09 18:51:54 crc kubenswrapper[4750]: I0309 18:51:54.678128 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" event={"ID":"ed71b380-e6d0-4f58-be59-c01a8de14ec5","Type":"ContainerStarted","Data":"0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59"} Mar 09 18:51:54 crc kubenswrapper[4750]: I0309 18:51:54.679437 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:51:54 crc kubenswrapper[4750]: I0309 18:51:54.717678 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" podStartSLOduration=3.717622319 podStartE2EDuration="3.717622319s" podCreationTimestamp="2026-03-09 18:51:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:51:54.70774501 +0000 UTC m=+1596.050217458" watchObservedRunningTime="2026-03-09 18:51:54.717622319 +0000 UTC m=+1596.060094747" Mar 09 18:51:55 crc kubenswrapper[4750]: I0309 18:51:55.394419 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3f3d445-e629-4668-a6f7-d8b44406e823" path="/var/lib/kubelet/pods/e3f3d445-e629-4668-a6f7-d8b44406e823/volumes" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.135689 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551372-tbfv8"] Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.137916 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.143086 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.145400 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.146301 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.163916 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551372-tbfv8"] Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.254583 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s8qd\" (UniqueName: \"kubernetes.io/projected/e1499c34-8025-4162-a499-662676c5e447-kube-api-access-2s8qd\") pod \"auto-csr-approver-29551372-tbfv8\" (UID: \"e1499c34-8025-4162-a499-662676c5e447\") " pod="openshift-infra/auto-csr-approver-29551372-tbfv8" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.357835 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s8qd\" (UniqueName: \"kubernetes.io/projected/e1499c34-8025-4162-a499-662676c5e447-kube-api-access-2s8qd\") pod \"auto-csr-approver-29551372-tbfv8\" (UID: \"e1499c34-8025-4162-a499-662676c5e447\") " pod="openshift-infra/auto-csr-approver-29551372-tbfv8" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.383028 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s8qd\" (UniqueName: \"kubernetes.io/projected/e1499c34-8025-4162-a499-662676c5e447-kube-api-access-2s8qd\") pod \"auto-csr-approver-29551372-tbfv8\" (UID: \"e1499c34-8025-4162-a499-662676c5e447\") " pod="openshift-infra/auto-csr-approver-29551372-tbfv8" Mar 09 18:52:00 crc kubenswrapper[4750]: I0309 18:52:00.471236 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" Mar 09 18:52:01 crc kubenswrapper[4750]: W0309 18:52:01.016332 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1499c34_8025_4162_a499_662676c5e447.slice/crio-1cf0d8f07763d25e3dbffb88b4ff5a15e5801d8de4a2004d33dbd3e3cc4d59bc WatchSource:0}: Error finding container 1cf0d8f07763d25e3dbffb88b4ff5a15e5801d8de4a2004d33dbd3e3cc4d59bc: Status 404 returned error can't find the container with id 1cf0d8f07763d25e3dbffb88b4ff5a15e5801d8de4a2004d33dbd3e3cc4d59bc Mar 09 18:52:01 crc kubenswrapper[4750]: I0309 18:52:01.016819 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551372-tbfv8"] Mar 09 18:52:01 crc kubenswrapper[4750]: I0309 18:52:01.020057 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 18:52:01 crc kubenswrapper[4750]: I0309 18:52:01.776485 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" event={"ID":"e1499c34-8025-4162-a499-662676c5e447","Type":"ContainerStarted","Data":"1cf0d8f07763d25e3dbffb88b4ff5a15e5801d8de4a2004d33dbd3e3cc4d59bc"} Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.233826 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.311192 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89669fc97-cr5m9"] Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.311513 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" podUID="46f02dee-def0-448a-8813-cf4e9083fec2" containerName="dnsmasq-dns" containerID="cri-o://bf7e8ab45df770e7d117f6e00b482d9df09dcc1557d1974dd5f480e7beda8062" gracePeriod=10 Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.478488 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69685b9687-h4kfk"] Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.480354 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.502182 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69685b9687-h4kfk"] Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.620436 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-ovsdbserver-nb\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.620841 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4msp\" (UniqueName: \"kubernetes.io/projected/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-kube-api-access-v4msp\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.620904 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-dns-swift-storage-0\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.620967 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-dns-svc\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.621000 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-openstack-edpm-ipam\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.621025 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-ovsdbserver-sb\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.621083 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-config\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.725731 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-dns-svc\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.725875 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-openstack-edpm-ipam\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.725928 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-ovsdbserver-sb\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.725993 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-config\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.726128 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-ovsdbserver-nb\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.726161 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4msp\" (UniqueName: \"kubernetes.io/projected/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-kube-api-access-v4msp\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.726279 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-dns-swift-storage-0\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.727555 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-dns-swift-storage-0\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.728383 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-config\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.728387 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-ovsdbserver-sb\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.729117 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-ovsdbserver-nb\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.730175 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-dns-svc\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.730429 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-openstack-edpm-ipam\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.752693 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4msp\" (UniqueName: \"kubernetes.io/projected/19a73cb2-d4f6-41fa-a9ee-8a949bd75c73-kube-api-access-v4msp\") pod \"dnsmasq-dns-69685b9687-h4kfk\" (UID: \"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73\") " pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.806079 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" event={"ID":"e1499c34-8025-4162-a499-662676c5e447","Type":"ContainerStarted","Data":"4dce597ebad76d251e228f192c4d14f53450ecec2148f3ef5d696c5786590fe4"} Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.819731 4750 generic.go:334] "Generic (PLEG): container finished" podID="46f02dee-def0-448a-8813-cf4e9083fec2" containerID="bf7e8ab45df770e7d117f6e00b482d9df09dcc1557d1974dd5f480e7beda8062" exitCode=0 Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.819783 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" event={"ID":"46f02dee-def0-448a-8813-cf4e9083fec2","Type":"ContainerDied","Data":"bf7e8ab45df770e7d117f6e00b482d9df09dcc1557d1974dd5f480e7beda8062"} Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.825653 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" podStartSLOduration=1.831209562 podStartE2EDuration="2.825595605s" podCreationTimestamp="2026-03-09 18:52:00 +0000 UTC" firstStartedPulling="2026-03-09 18:52:01.01976114 +0000 UTC m=+1602.362233538" lastFinishedPulling="2026-03-09 18:52:02.014147183 +0000 UTC m=+1603.356619581" observedRunningTime="2026-03-09 18:52:02.819512238 +0000 UTC m=+1604.161984636" watchObservedRunningTime="2026-03-09 18:52:02.825595605 +0000 UTC m=+1604.168068003" Mar 09 18:52:02 crc kubenswrapper[4750]: I0309 18:52:02.837334 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.006343 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.139114 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-sb\") pod \"46f02dee-def0-448a-8813-cf4e9083fec2\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.140495 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6wbg\" (UniqueName: \"kubernetes.io/projected/46f02dee-def0-448a-8813-cf4e9083fec2-kube-api-access-z6wbg\") pod \"46f02dee-def0-448a-8813-cf4e9083fec2\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.140601 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-nb\") pod \"46f02dee-def0-448a-8813-cf4e9083fec2\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.140661 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-svc\") pod \"46f02dee-def0-448a-8813-cf4e9083fec2\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.140706 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-config\") pod \"46f02dee-def0-448a-8813-cf4e9083fec2\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.140783 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-swift-storage-0\") pod \"46f02dee-def0-448a-8813-cf4e9083fec2\" (UID: \"46f02dee-def0-448a-8813-cf4e9083fec2\") " Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.168729 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f02dee-def0-448a-8813-cf4e9083fec2-kube-api-access-z6wbg" (OuterVolumeSpecName: "kube-api-access-z6wbg") pod "46f02dee-def0-448a-8813-cf4e9083fec2" (UID: "46f02dee-def0-448a-8813-cf4e9083fec2"). InnerVolumeSpecName "kube-api-access-z6wbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.248141 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6wbg\" (UniqueName: \"kubernetes.io/projected/46f02dee-def0-448a-8813-cf4e9083fec2-kube-api-access-z6wbg\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.274435 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69685b9687-h4kfk"] Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.326153 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-config" (OuterVolumeSpecName: "config") pod "46f02dee-def0-448a-8813-cf4e9083fec2" (UID: "46f02dee-def0-448a-8813-cf4e9083fec2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.351030 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.368227 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "46f02dee-def0-448a-8813-cf4e9083fec2" (UID: "46f02dee-def0-448a-8813-cf4e9083fec2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.418492 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "46f02dee-def0-448a-8813-cf4e9083fec2" (UID: "46f02dee-def0-448a-8813-cf4e9083fec2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.419070 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "46f02dee-def0-448a-8813-cf4e9083fec2" (UID: "46f02dee-def0-448a-8813-cf4e9083fec2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.426773 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "46f02dee-def0-448a-8813-cf4e9083fec2" (UID: "46f02dee-def0-448a-8813-cf4e9083fec2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.454509 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.454544 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.454560 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.454570 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46f02dee-def0-448a-8813-cf4e9083fec2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.845154 4750 generic.go:334] "Generic (PLEG): container finished" podID="19a73cb2-d4f6-41fa-a9ee-8a949bd75c73" containerID="c6e38d5255b790757fe6012265300856c7303ebfc782b95ffe6946dc0c9968fe" exitCode=0 Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.845241 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" event={"ID":"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73","Type":"ContainerDied","Data":"c6e38d5255b790757fe6012265300856c7303ebfc782b95ffe6946dc0c9968fe"} Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.846323 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" event={"ID":"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73","Type":"ContainerStarted","Data":"6d3e74340b07a834c18cdc400ffe4122fcb1d86af0e0fd6e1c963ec223402da4"} Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.855213 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.855680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89669fc97-cr5m9" event={"ID":"46f02dee-def0-448a-8813-cf4e9083fec2","Type":"ContainerDied","Data":"8fec15ec5a11ca324f9f2d6aeedcc4013248ae4d53de78e4bab9c42b45271497"} Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.855754 4750 scope.go:117] "RemoveContainer" containerID="bf7e8ab45df770e7d117f6e00b482d9df09dcc1557d1974dd5f480e7beda8062" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.860019 4750 generic.go:334] "Generic (PLEG): container finished" podID="e1499c34-8025-4162-a499-662676c5e447" containerID="4dce597ebad76d251e228f192c4d14f53450ecec2148f3ef5d696c5786590fe4" exitCode=0 Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.860058 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" event={"ID":"e1499c34-8025-4162-a499-662676c5e447","Type":"ContainerDied","Data":"4dce597ebad76d251e228f192c4d14f53450ecec2148f3ef5d696c5786590fe4"} Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.928085 4750 scope.go:117] "RemoveContainer" containerID="e87ffa1b9c2af86bc4fcc1571081142655397fce9bae64cdda8a84223b111c11" Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.951232 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89669fc97-cr5m9"] Mar 09 18:52:03 crc kubenswrapper[4750]: I0309 18:52:03.960682 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89669fc97-cr5m9"] Mar 09 18:52:04 crc kubenswrapper[4750]: I0309 18:52:04.873350 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" event={"ID":"19a73cb2-d4f6-41fa-a9ee-8a949bd75c73","Type":"ContainerStarted","Data":"8263a712a2b32e3b7fc74b44005cdcb547e529a00d80095591801a8ff2348eaa"} Mar 09 18:52:04 crc kubenswrapper[4750]: I0309 18:52:04.874949 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:04 crc kubenswrapper[4750]: I0309 18:52:04.916010 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" podStartSLOduration=2.9159911899999997 podStartE2EDuration="2.91599119s" podCreationTimestamp="2026-03-09 18:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:52:04.895643316 +0000 UTC m=+1606.238115734" watchObservedRunningTime="2026-03-09 18:52:04.91599119 +0000 UTC m=+1606.258463588" Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.305016 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.385194 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f02dee-def0-448a-8813-cf4e9083fec2" path="/var/lib/kubelet/pods/46f02dee-def0-448a-8813-cf4e9083fec2/volumes" Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.499877 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s8qd\" (UniqueName: \"kubernetes.io/projected/e1499c34-8025-4162-a499-662676c5e447-kube-api-access-2s8qd\") pod \"e1499c34-8025-4162-a499-662676c5e447\" (UID: \"e1499c34-8025-4162-a499-662676c5e447\") " Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.514543 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1499c34-8025-4162-a499-662676c5e447-kube-api-access-2s8qd" (OuterVolumeSpecName: "kube-api-access-2s8qd") pod "e1499c34-8025-4162-a499-662676c5e447" (UID: "e1499c34-8025-4162-a499-662676c5e447"). InnerVolumeSpecName "kube-api-access-2s8qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.602588 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s8qd\" (UniqueName: \"kubernetes.io/projected/e1499c34-8025-4162-a499-662676c5e447-kube-api-access-2s8qd\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.896979 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.897394 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551372-tbfv8" event={"ID":"e1499c34-8025-4162-a499-662676c5e447","Type":"ContainerDied","Data":"1cf0d8f07763d25e3dbffb88b4ff5a15e5801d8de4a2004d33dbd3e3cc4d59bc"} Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.897441 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf0d8f07763d25e3dbffb88b4ff5a15e5801d8de4a2004d33dbd3e3cc4d59bc" Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.931703 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551366-ds24r"] Mar 09 18:52:05 crc kubenswrapper[4750]: I0309 18:52:05.946548 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551366-ds24r"] Mar 09 18:52:07 crc kubenswrapper[4750]: I0309 18:52:07.394102 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7567fbed-2455-4d79-86fc-296893f49e2d" path="/var/lib/kubelet/pods/7567fbed-2455-4d79-86fc-296893f49e2d/volumes" Mar 09 18:52:12 crc kubenswrapper[4750]: I0309 18:52:12.841053 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69685b9687-h4kfk" Mar 09 18:52:12 crc kubenswrapper[4750]: I0309 18:52:12.937707 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56d7fdf8b5-4kbkb"] Mar 09 18:52:12 crc kubenswrapper[4750]: I0309 18:52:12.938158 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" podUID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerName="dnsmasq-dns" containerID="cri-o://0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59" gracePeriod=10 Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.475302 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.592129 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-nb\") pod \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.592213 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-swift-storage-0\") pod \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.592239 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7bns\" (UniqueName: \"kubernetes.io/projected/ed71b380-e6d0-4f58-be59-c01a8de14ec5-kube-api-access-n7bns\") pod \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.592356 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-svc\") pod \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.592456 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-sb\") pod \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.592529 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-openstack-edpm-ipam\") pod \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.592592 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-config\") pod \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\" (UID: \"ed71b380-e6d0-4f58-be59-c01a8de14ec5\") " Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.607004 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed71b380-e6d0-4f58-be59-c01a8de14ec5-kube-api-access-n7bns" (OuterVolumeSpecName: "kube-api-access-n7bns") pod "ed71b380-e6d0-4f58-be59-c01a8de14ec5" (UID: "ed71b380-e6d0-4f58-be59-c01a8de14ec5"). InnerVolumeSpecName "kube-api-access-n7bns". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.650194 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed71b380-e6d0-4f58-be59-c01a8de14ec5" (UID: "ed71b380-e6d0-4f58-be59-c01a8de14ec5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.652205 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ed71b380-e6d0-4f58-be59-c01a8de14ec5" (UID: "ed71b380-e6d0-4f58-be59-c01a8de14ec5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.654124 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed71b380-e6d0-4f58-be59-c01a8de14ec5" (UID: "ed71b380-e6d0-4f58-be59-c01a8de14ec5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.662181 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ed71b380-e6d0-4f58-be59-c01a8de14ec5" (UID: "ed71b380-e6d0-4f58-be59-c01a8de14ec5"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.663919 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-config" (OuterVolumeSpecName: "config") pod "ed71b380-e6d0-4f58-be59-c01a8de14ec5" (UID: "ed71b380-e6d0-4f58-be59-c01a8de14ec5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.683375 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed71b380-e6d0-4f58-be59-c01a8de14ec5" (UID: "ed71b380-e6d0-4f58-be59-c01a8de14ec5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.695086 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.695466 4750 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.695493 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-config\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.695588 4750 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.695602 4750 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.695611 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7bns\" (UniqueName: \"kubernetes.io/projected/ed71b380-e6d0-4f58-be59-c01a8de14ec5-kube-api-access-n7bns\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.695621 4750 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed71b380-e6d0-4f58-be59-c01a8de14ec5-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.990680 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.990623 4750 generic.go:334] "Generic (PLEG): container finished" podID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerID="0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59" exitCode=0 Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.990717 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" event={"ID":"ed71b380-e6d0-4f58-be59-c01a8de14ec5","Type":"ContainerDied","Data":"0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59"} Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.991732 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56d7fdf8b5-4kbkb" event={"ID":"ed71b380-e6d0-4f58-be59-c01a8de14ec5","Type":"ContainerDied","Data":"cdd52e366ed5c414c71b73a7a44f1dcfa8f5ea59cfd5b0106f2c879b027e7f42"} Mar 09 18:52:13 crc kubenswrapper[4750]: I0309 18:52:13.991772 4750 scope.go:117] "RemoveContainer" containerID="0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59" Mar 09 18:52:14 crc kubenswrapper[4750]: I0309 18:52:14.025606 4750 scope.go:117] "RemoveContainer" containerID="4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3" Mar 09 18:52:14 crc kubenswrapper[4750]: I0309 18:52:14.032579 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56d7fdf8b5-4kbkb"] Mar 09 18:52:14 crc kubenswrapper[4750]: I0309 18:52:14.043097 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56d7fdf8b5-4kbkb"] Mar 09 18:52:14 crc kubenswrapper[4750]: I0309 18:52:14.055010 4750 scope.go:117] "RemoveContainer" containerID="0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59" Mar 09 18:52:14 crc kubenswrapper[4750]: E0309 18:52:14.055452 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59\": container with ID starting with 0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59 not found: ID does not exist" containerID="0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59" Mar 09 18:52:14 crc kubenswrapper[4750]: I0309 18:52:14.055548 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59"} err="failed to get container status \"0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59\": rpc error: code = NotFound desc = could not find container \"0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59\": container with ID starting with 0c93c9ee31a42b27db0a4613912c17c636ef53334e9ef175330513cc33bdce59 not found: ID does not exist" Mar 09 18:52:14 crc kubenswrapper[4750]: I0309 18:52:14.055626 4750 scope.go:117] "RemoveContainer" containerID="4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3" Mar 09 18:52:14 crc kubenswrapper[4750]: E0309 18:52:14.055988 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3\": container with ID starting with 4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3 not found: ID does not exist" containerID="4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3" Mar 09 18:52:14 crc kubenswrapper[4750]: I0309 18:52:14.056071 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3"} err="failed to get container status \"4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3\": rpc error: code = NotFound desc = could not find container \"4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3\": container with ID starting with 4a22e0d37142b7583db040bd90e61cc451745058fe120d0d3d558043c548dbe3 not found: ID does not exist" Mar 09 18:52:15 crc kubenswrapper[4750]: I0309 18:52:15.384928 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" path="/var/lib/kubelet/pods/ed71b380-e6d0-4f58-be59-c01a8de14ec5/volumes" Mar 09 18:52:19 crc kubenswrapper[4750]: I0309 18:52:19.056368 4750 generic.go:334] "Generic (PLEG): container finished" podID="b4103fad-b50b-450d-99c0-38a505bd800a" containerID="1358bbd2528c010830a89b315b8dd22a7801a0040a42c2a6fbca19f26fee7d45" exitCode=0 Mar 09 18:52:19 crc kubenswrapper[4750]: I0309 18:52:19.056476 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4103fad-b50b-450d-99c0-38a505bd800a","Type":"ContainerDied","Data":"1358bbd2528c010830a89b315b8dd22a7801a0040a42c2a6fbca19f26fee7d45"} Mar 09 18:52:19 crc kubenswrapper[4750]: I0309 18:52:19.058929 4750 generic.go:334] "Generic (PLEG): container finished" podID="14f2384b-45f7-45ca-ac90-429fc2f69bbb" containerID="99bef4fd42c532835b9aa16a998190da9547d9c841d5438d0c51e7a1a7992b8c" exitCode=0 Mar 09 18:52:19 crc kubenswrapper[4750]: I0309 18:52:19.058958 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"14f2384b-45f7-45ca-ac90-429fc2f69bbb","Type":"ContainerDied","Data":"99bef4fd42c532835b9aa16a998190da9547d9c841d5438d0c51e7a1a7992b8c"} Mar 09 18:52:20 crc kubenswrapper[4750]: I0309 18:52:20.077907 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4103fad-b50b-450d-99c0-38a505bd800a","Type":"ContainerStarted","Data":"6e0c78c296f8f6cd223538fbc9163bf219490696e1571bc493336e75b116c889"} Mar 09 18:52:20 crc kubenswrapper[4750]: I0309 18:52:20.078717 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:52:20 crc kubenswrapper[4750]: I0309 18:52:20.081236 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"14f2384b-45f7-45ca-ac90-429fc2f69bbb","Type":"ContainerStarted","Data":"c02a1a04e01285aad8a10e855fe50eca3a9348e09706e1dc75a0fb6a5eebd583"} Mar 09 18:52:20 crc kubenswrapper[4750]: I0309 18:52:20.081549 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 09 18:52:20 crc kubenswrapper[4750]: I0309 18:52:20.111090 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.111062425 podStartE2EDuration="37.111062425s" podCreationTimestamp="2026-03-09 18:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:52:20.105381181 +0000 UTC m=+1621.447853609" watchObservedRunningTime="2026-03-09 18:52:20.111062425 +0000 UTC m=+1621.453534833" Mar 09 18:52:20 crc kubenswrapper[4750]: I0309 18:52:20.138850 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.138827502 podStartE2EDuration="37.138827502s" podCreationTimestamp="2026-03-09 18:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 18:52:20.1372949 +0000 UTC m=+1621.479767298" watchObservedRunningTime="2026-03-09 18:52:20.138827502 +0000 UTC m=+1621.481299900" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.659991 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw"] Mar 09 18:52:30 crc kubenswrapper[4750]: E0309 18:52:30.661112 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f02dee-def0-448a-8813-cf4e9083fec2" containerName="dnsmasq-dns" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661125 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f02dee-def0-448a-8813-cf4e9083fec2" containerName="dnsmasq-dns" Mar 09 18:52:30 crc kubenswrapper[4750]: E0309 18:52:30.661143 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1499c34-8025-4162-a499-662676c5e447" containerName="oc" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661149 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1499c34-8025-4162-a499-662676c5e447" containerName="oc" Mar 09 18:52:30 crc kubenswrapper[4750]: E0309 18:52:30.661184 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f02dee-def0-448a-8813-cf4e9083fec2" containerName="init" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661190 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f02dee-def0-448a-8813-cf4e9083fec2" containerName="init" Mar 09 18:52:30 crc kubenswrapper[4750]: E0309 18:52:30.661200 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerName="init" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661205 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerName="init" Mar 09 18:52:30 crc kubenswrapper[4750]: E0309 18:52:30.661218 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerName="dnsmasq-dns" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661224 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerName="dnsmasq-dns" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661413 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1499c34-8025-4162-a499-662676c5e447" containerName="oc" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661423 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f02dee-def0-448a-8813-cf4e9083fec2" containerName="dnsmasq-dns" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.661436 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed71b380-e6d0-4f58-be59-c01a8de14ec5" containerName="dnsmasq-dns" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.662167 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.674580 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.675258 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.675878 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.676005 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.687236 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw"] Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.775228 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw6ks\" (UniqueName: \"kubernetes.io/projected/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-kube-api-access-rw6ks\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.775302 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.775387 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.775678 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.877555 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.877709 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.877798 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw6ks\" (UniqueName: \"kubernetes.io/projected/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-kube-api-access-rw6ks\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.877825 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.885192 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.885511 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.886137 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:30 crc kubenswrapper[4750]: I0309 18:52:30.898895 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw6ks\" (UniqueName: \"kubernetes.io/projected/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-kube-api-access-rw6ks\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:31 crc kubenswrapper[4750]: I0309 18:52:31.037343 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:31 crc kubenswrapper[4750]: I0309 18:52:31.912189 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw"] Mar 09 18:52:32 crc kubenswrapper[4750]: I0309 18:52:32.218874 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" event={"ID":"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6","Type":"ContainerStarted","Data":"5c0e81e66cd846e86e8f7e8365e083d5f1b2d00b538a184c8343e232254f0040"} Mar 09 18:52:34 crc kubenswrapper[4750]: I0309 18:52:34.053360 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="14f2384b-45f7-45ca-ac90-429fc2f69bbb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.244:5671: connect: connection refused" Mar 09 18:52:34 crc kubenswrapper[4750]: I0309 18:52:34.319989 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b4103fad-b50b-450d-99c0-38a505bd800a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.243:5671: connect: connection refused" Mar 09 18:52:41 crc kubenswrapper[4750]: I0309 18:52:41.342133 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" event={"ID":"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6","Type":"ContainerStarted","Data":"958222efa791eca44a0c29dc30a72d80b9742a5e60f893300a2ff1dbb143a8da"} Mar 09 18:52:41 crc kubenswrapper[4750]: I0309 18:52:41.368498 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" podStartSLOduration=2.575490115 podStartE2EDuration="11.368470677s" podCreationTimestamp="2026-03-09 18:52:30 +0000 UTC" firstStartedPulling="2026-03-09 18:52:31.909781684 +0000 UTC m=+1633.252254122" lastFinishedPulling="2026-03-09 18:52:40.702762276 +0000 UTC m=+1642.045234684" observedRunningTime="2026-03-09 18:52:41.364368287 +0000 UTC m=+1642.706840685" watchObservedRunningTime="2026-03-09 18:52:41.368470677 +0000 UTC m=+1642.710943075" Mar 09 18:52:41 crc kubenswrapper[4750]: I0309 18:52:41.404058 4750 scope.go:117] "RemoveContainer" containerID="f719e033a147462c6c3db9db045bd1371c2b1a1f7a8cf0b4284eae84a2d5ecc2" Mar 09 18:52:41 crc kubenswrapper[4750]: I0309 18:52:41.475902 4750 scope.go:117] "RemoveContainer" containerID="1ed9d78873c42f484fd82ac83617e0adac66f1e76adc7b75b939bd6244566400" Mar 09 18:52:41 crc kubenswrapper[4750]: I0309 18:52:41.508519 4750 scope.go:117] "RemoveContainer" containerID="7e8b256fe6a419bbbdb8af0b37473cc0f0e8beb045d44fc31abc6733df215763" Mar 09 18:52:41 crc kubenswrapper[4750]: I0309 18:52:41.566820 4750 scope.go:117] "RemoveContainer" containerID="75ae4399a621829d965c2412052f8815980f3acb9857214f9823c0287999a261" Mar 09 18:52:41 crc kubenswrapper[4750]: I0309 18:52:41.608008 4750 scope.go:117] "RemoveContainer" containerID="49fbad12eb84b0f3d14f848f220cc82823fd7871d7abf0b485264235d15dc73f" Mar 09 18:52:44 crc kubenswrapper[4750]: I0309 18:52:44.053921 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 09 18:52:44 crc kubenswrapper[4750]: I0309 18:52:44.318864 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 09 18:52:52 crc kubenswrapper[4750]: I0309 18:52:52.505490 4750 generic.go:334] "Generic (PLEG): container finished" podID="dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" containerID="958222efa791eca44a0c29dc30a72d80b9742a5e60f893300a2ff1dbb143a8da" exitCode=0 Mar 09 18:52:52 crc kubenswrapper[4750]: I0309 18:52:52.505934 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" event={"ID":"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6","Type":"ContainerDied","Data":"958222efa791eca44a0c29dc30a72d80b9742a5e60f893300a2ff1dbb143a8da"} Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.013916 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.143337 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw6ks\" (UniqueName: \"kubernetes.io/projected/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-kube-api-access-rw6ks\") pod \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.143522 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-repo-setup-combined-ca-bundle\") pod \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.143557 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-ssh-key-openstack-edpm-ipam\") pod \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.143683 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-inventory\") pod \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\" (UID: \"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6\") " Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.149982 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" (UID: "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.150431 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-kube-api-access-rw6ks" (OuterVolumeSpecName: "kube-api-access-rw6ks") pod "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" (UID: "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6"). InnerVolumeSpecName "kube-api-access-rw6ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.173473 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" (UID: "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.175676 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-inventory" (OuterVolumeSpecName: "inventory") pod "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" (UID: "dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.246494 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.246554 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw6ks\" (UniqueName: \"kubernetes.io/projected/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-kube-api-access-rw6ks\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.246578 4750 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.246596 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.530906 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" event={"ID":"dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6","Type":"ContainerDied","Data":"5c0e81e66cd846e86e8f7e8365e083d5f1b2d00b538a184c8343e232254f0040"} Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.530948 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c0e81e66cd846e86e8f7e8365e083d5f1b2d00b538a184c8343e232254f0040" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.531005 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.624665 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct"] Mar 09 18:52:54 crc kubenswrapper[4750]: E0309 18:52:54.626653 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.626684 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.626949 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.631481 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.643060 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.643267 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.643270 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.643798 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.652658 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct"] Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.790942 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7vdv\" (UniqueName: \"kubernetes.io/projected/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-kube-api-access-w7vdv\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.791300 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.791518 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.893527 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.893705 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7vdv\" (UniqueName: \"kubernetes.io/projected/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-kube-api-access-w7vdv\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.893770 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.898183 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.898468 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.925130 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7vdv\" (UniqueName: \"kubernetes.io/projected/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-kube-api-access-w7vdv\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tdfct\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:54 crc kubenswrapper[4750]: I0309 18:52:54.955559 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:52:55 crc kubenswrapper[4750]: I0309 18:52:55.543613 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct"] Mar 09 18:52:55 crc kubenswrapper[4750]: W0309 18:52:55.549129 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4dc3075_1b7c_4076_9beb_f7e5471c2dbe.slice/crio-3fd4a5404cfeb2e8f6bf1f57be1b9dde1d64b9b27ff72abcaeeca7010f108050 WatchSource:0}: Error finding container 3fd4a5404cfeb2e8f6bf1f57be1b9dde1d64b9b27ff72abcaeeca7010f108050: Status 404 returned error can't find the container with id 3fd4a5404cfeb2e8f6bf1f57be1b9dde1d64b9b27ff72abcaeeca7010f108050 Mar 09 18:52:56 crc kubenswrapper[4750]: I0309 18:52:56.555838 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" event={"ID":"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe","Type":"ContainerStarted","Data":"db8829477624fb0959290f7dd537f0c3bd3932b14ea66f6c32c0379ea8cd705a"} Mar 09 18:52:56 crc kubenswrapper[4750]: I0309 18:52:56.556223 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" event={"ID":"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe","Type":"ContainerStarted","Data":"3fd4a5404cfeb2e8f6bf1f57be1b9dde1d64b9b27ff72abcaeeca7010f108050"} Mar 09 18:52:56 crc kubenswrapper[4750]: I0309 18:52:56.583309 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" podStartSLOduration=2.164242367 podStartE2EDuration="2.58329015s" podCreationTimestamp="2026-03-09 18:52:54 +0000 UTC" firstStartedPulling="2026-03-09 18:52:55.552377432 +0000 UTC m=+1656.894849830" lastFinishedPulling="2026-03-09 18:52:55.971425185 +0000 UTC m=+1657.313897613" observedRunningTime="2026-03-09 18:52:56.575986812 +0000 UTC m=+1657.918459210" watchObservedRunningTime="2026-03-09 18:52:56.58329015 +0000 UTC m=+1657.925762538" Mar 09 18:52:59 crc kubenswrapper[4750]: I0309 18:52:59.607264 4750 generic.go:334] "Generic (PLEG): container finished" podID="c4dc3075-1b7c-4076-9beb-f7e5471c2dbe" containerID="db8829477624fb0959290f7dd537f0c3bd3932b14ea66f6c32c0379ea8cd705a" exitCode=0 Mar 09 18:52:59 crc kubenswrapper[4750]: I0309 18:52:59.607376 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" event={"ID":"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe","Type":"ContainerDied","Data":"db8829477624fb0959290f7dd537f0c3bd3932b14ea66f6c32c0379ea8cd705a"} Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.119378 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.163951 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7vdv\" (UniqueName: \"kubernetes.io/projected/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-kube-api-access-w7vdv\") pod \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.164003 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-inventory\") pod \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.164136 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-ssh-key-openstack-edpm-ipam\") pod \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\" (UID: \"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe\") " Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.196381 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-kube-api-access-w7vdv" (OuterVolumeSpecName: "kube-api-access-w7vdv") pod "c4dc3075-1b7c-4076-9beb-f7e5471c2dbe" (UID: "c4dc3075-1b7c-4076-9beb-f7e5471c2dbe"). InnerVolumeSpecName "kube-api-access-w7vdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.234686 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c4dc3075-1b7c-4076-9beb-f7e5471c2dbe" (UID: "c4dc3075-1b7c-4076-9beb-f7e5471c2dbe"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.245458 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-inventory" (OuterVolumeSpecName: "inventory") pod "c4dc3075-1b7c-4076-9beb-f7e5471c2dbe" (UID: "c4dc3075-1b7c-4076-9beb-f7e5471c2dbe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.265328 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.265359 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7vdv\" (UniqueName: \"kubernetes.io/projected/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-kube-api-access-w7vdv\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.265368 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4dc3075-1b7c-4076-9beb-f7e5471c2dbe-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.633172 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" event={"ID":"c4dc3075-1b7c-4076-9beb-f7e5471c2dbe","Type":"ContainerDied","Data":"3fd4a5404cfeb2e8f6bf1f57be1b9dde1d64b9b27ff72abcaeeca7010f108050"} Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.633232 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fd4a5404cfeb2e8f6bf1f57be1b9dde1d64b9b27ff72abcaeeca7010f108050" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.633243 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tdfct" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.715954 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4"] Mar 09 18:53:01 crc kubenswrapper[4750]: E0309 18:53:01.716454 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4dc3075-1b7c-4076-9beb-f7e5471c2dbe" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.716479 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4dc3075-1b7c-4076-9beb-f7e5471c2dbe" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.716750 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4dc3075-1b7c-4076-9beb-f7e5471c2dbe" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.717616 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.719681 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.719956 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.719959 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.720134 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.733230 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4"] Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.773759 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.773854 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.774052 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t495x\" (UniqueName: \"kubernetes.io/projected/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-kube-api-access-t495x\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.774115 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.874891 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t495x\" (UniqueName: \"kubernetes.io/projected/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-kube-api-access-t495x\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.874948 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.875057 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.875093 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.879088 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.879176 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.879991 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:01 crc kubenswrapper[4750]: I0309 18:53:01.903019 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t495x\" (UniqueName: \"kubernetes.io/projected/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-kube-api-access-t495x\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:02 crc kubenswrapper[4750]: I0309 18:53:02.039277 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:53:02 crc kubenswrapper[4750]: W0309 18:53:02.600454 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedf12ec8_96ec_4d8e_a9f5_bbdd0cdcc4ee.slice/crio-3bb2df63f15b16956ab97e75979433db8aaab92aca16183a0ffbd0de4f305f9a WatchSource:0}: Error finding container 3bb2df63f15b16956ab97e75979433db8aaab92aca16183a0ffbd0de4f305f9a: Status 404 returned error can't find the container with id 3bb2df63f15b16956ab97e75979433db8aaab92aca16183a0ffbd0de4f305f9a Mar 09 18:53:02 crc kubenswrapper[4750]: I0309 18:53:02.635789 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4"] Mar 09 18:53:02 crc kubenswrapper[4750]: I0309 18:53:02.643812 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" event={"ID":"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee","Type":"ContainerStarted","Data":"3bb2df63f15b16956ab97e75979433db8aaab92aca16183a0ffbd0de4f305f9a"} Mar 09 18:53:03 crc kubenswrapper[4750]: I0309 18:53:03.658903 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" event={"ID":"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee","Type":"ContainerStarted","Data":"8b87065ef1e47a9957f2ddc4056e605f9227718859027be5b5520c09f2fa02a0"} Mar 09 18:53:03 crc kubenswrapper[4750]: I0309 18:53:03.689476 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" podStartSLOduration=2.201653672 podStartE2EDuration="2.689456889s" podCreationTimestamp="2026-03-09 18:53:01 +0000 UTC" firstStartedPulling="2026-03-09 18:53:02.603973454 +0000 UTC m=+1663.946445862" lastFinishedPulling="2026-03-09 18:53:03.091776681 +0000 UTC m=+1664.434249079" observedRunningTime="2026-03-09 18:53:03.681359279 +0000 UTC m=+1665.023831717" watchObservedRunningTime="2026-03-09 18:53:03.689456889 +0000 UTC m=+1665.031929287" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.520082 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b6z5n"] Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.523437 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.534564 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b6z5n"] Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.560046 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f59jv\" (UniqueName: \"kubernetes.io/projected/a501006f-283a-4dc1-8498-733c909bdae0-kube-api-access-f59jv\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.560309 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-utilities\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.560381 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-catalog-content\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.669469 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-utilities\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.669526 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-catalog-content\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.669593 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f59jv\" (UniqueName: \"kubernetes.io/projected/a501006f-283a-4dc1-8498-733c909bdae0-kube-api-access-f59jv\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.670086 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-utilities\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.670584 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-catalog-content\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.694577 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f59jv\" (UniqueName: \"kubernetes.io/projected/a501006f-283a-4dc1-8498-733c909bdae0-kube-api-access-f59jv\") pod \"community-operators-b6z5n\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:09 crc kubenswrapper[4750]: I0309 18:53:09.866107 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:10 crc kubenswrapper[4750]: I0309 18:53:10.481195 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b6z5n"] Mar 09 18:53:10 crc kubenswrapper[4750]: W0309 18:53:10.483395 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda501006f_283a_4dc1_8498_733c909bdae0.slice/crio-237fcb1e011b481d42164c0ecc0db93990ee262d4d39a1c453fd68922af079ff WatchSource:0}: Error finding container 237fcb1e011b481d42164c0ecc0db93990ee262d4d39a1c453fd68922af079ff: Status 404 returned error can't find the container with id 237fcb1e011b481d42164c0ecc0db93990ee262d4d39a1c453fd68922af079ff Mar 09 18:53:10 crc kubenswrapper[4750]: I0309 18:53:10.742136 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerStarted","Data":"c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07"} Mar 09 18:53:10 crc kubenswrapper[4750]: I0309 18:53:10.742543 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerStarted","Data":"237fcb1e011b481d42164c0ecc0db93990ee262d4d39a1c453fd68922af079ff"} Mar 09 18:53:11 crc kubenswrapper[4750]: I0309 18:53:11.754472 4750 generic.go:334] "Generic (PLEG): container finished" podID="a501006f-283a-4dc1-8498-733c909bdae0" containerID="c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07" exitCode=0 Mar 09 18:53:11 crc kubenswrapper[4750]: I0309 18:53:11.754532 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerDied","Data":"c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07"} Mar 09 18:53:12 crc kubenswrapper[4750]: I0309 18:53:12.768108 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerStarted","Data":"35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10"} Mar 09 18:53:14 crc kubenswrapper[4750]: I0309 18:53:14.799230 4750 generic.go:334] "Generic (PLEG): container finished" podID="a501006f-283a-4dc1-8498-733c909bdae0" containerID="35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10" exitCode=0 Mar 09 18:53:14 crc kubenswrapper[4750]: I0309 18:53:14.799388 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerDied","Data":"35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10"} Mar 09 18:53:15 crc kubenswrapper[4750]: I0309 18:53:15.813038 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerStarted","Data":"a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1"} Mar 09 18:53:15 crc kubenswrapper[4750]: I0309 18:53:15.838929 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b6z5n" podStartSLOduration=3.207481054 podStartE2EDuration="6.838904462s" podCreationTimestamp="2026-03-09 18:53:09 +0000 UTC" firstStartedPulling="2026-03-09 18:53:11.758603377 +0000 UTC m=+1673.101075775" lastFinishedPulling="2026-03-09 18:53:15.390026775 +0000 UTC m=+1676.732499183" observedRunningTime="2026-03-09 18:53:15.83300532 +0000 UTC m=+1677.175477738" watchObservedRunningTime="2026-03-09 18:53:15.838904462 +0000 UTC m=+1677.181376860" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.290800 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6kdxn"] Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.293063 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.313349 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6kdxn"] Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.382259 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-catalog-content\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.382341 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwp9p\" (UniqueName: \"kubernetes.io/projected/622ca8d5-72d9-4fce-ad90-9f46ce680820-kube-api-access-cwp9p\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.382484 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-utilities\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.484284 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-catalog-content\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.484355 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwp9p\" (UniqueName: \"kubernetes.io/projected/622ca8d5-72d9-4fce-ad90-9f46ce680820-kube-api-access-cwp9p\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.484430 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-utilities\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.485053 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-utilities\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.485056 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-catalog-content\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.513579 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwp9p\" (UniqueName: \"kubernetes.io/projected/622ca8d5-72d9-4fce-ad90-9f46ce680820-kube-api-access-cwp9p\") pod \"certified-operators-6kdxn\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:16 crc kubenswrapper[4750]: I0309 18:53:16.621737 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:17 crc kubenswrapper[4750]: I0309 18:53:17.221923 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6kdxn"] Mar 09 18:53:17 crc kubenswrapper[4750]: I0309 18:53:17.847155 4750 generic.go:334] "Generic (PLEG): container finished" podID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerID="fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b" exitCode=0 Mar 09 18:53:17 crc kubenswrapper[4750]: I0309 18:53:17.847483 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kdxn" event={"ID":"622ca8d5-72d9-4fce-ad90-9f46ce680820","Type":"ContainerDied","Data":"fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b"} Mar 09 18:53:17 crc kubenswrapper[4750]: I0309 18:53:17.847512 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kdxn" event={"ID":"622ca8d5-72d9-4fce-ad90-9f46ce680820","Type":"ContainerStarted","Data":"5d62135ec060099c5dff7f3503b0ca4316fe6a03b7c36d93f1834af59e4ae711"} Mar 09 18:53:18 crc kubenswrapper[4750]: I0309 18:53:18.881724 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kdxn" event={"ID":"622ca8d5-72d9-4fce-ad90-9f46ce680820","Type":"ContainerStarted","Data":"7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761"} Mar 09 18:53:19 crc kubenswrapper[4750]: I0309 18:53:19.867202 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:19 crc kubenswrapper[4750]: I0309 18:53:19.867596 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:19 crc kubenswrapper[4750]: I0309 18:53:19.930976 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:21 crc kubenswrapper[4750]: I0309 18:53:21.744127 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:53:21 crc kubenswrapper[4750]: I0309 18:53:21.744585 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:53:21 crc kubenswrapper[4750]: I0309 18:53:21.935775 4750 generic.go:334] "Generic (PLEG): container finished" podID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerID="7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761" exitCode=0 Mar 09 18:53:21 crc kubenswrapper[4750]: I0309 18:53:21.935836 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kdxn" event={"ID":"622ca8d5-72d9-4fce-ad90-9f46ce680820","Type":"ContainerDied","Data":"7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761"} Mar 09 18:53:22 crc kubenswrapper[4750]: I0309 18:53:22.950938 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kdxn" event={"ID":"622ca8d5-72d9-4fce-ad90-9f46ce680820","Type":"ContainerStarted","Data":"14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534"} Mar 09 18:53:22 crc kubenswrapper[4750]: I0309 18:53:22.977463 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6kdxn" podStartSLOduration=2.400248764 podStartE2EDuration="6.977435922s" podCreationTimestamp="2026-03-09 18:53:16 +0000 UTC" firstStartedPulling="2026-03-09 18:53:17.851866568 +0000 UTC m=+1679.194338976" lastFinishedPulling="2026-03-09 18:53:22.429053726 +0000 UTC m=+1683.771526134" observedRunningTime="2026-03-09 18:53:22.973986699 +0000 UTC m=+1684.316459107" watchObservedRunningTime="2026-03-09 18:53:22.977435922 +0000 UTC m=+1684.319908330" Mar 09 18:53:26 crc kubenswrapper[4750]: I0309 18:53:26.622597 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:26 crc kubenswrapper[4750]: I0309 18:53:26.623288 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:26 crc kubenswrapper[4750]: I0309 18:53:26.675492 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.522052 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jvqk6"] Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.525108 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.536697 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvqk6"] Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.575283 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-catalog-content\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.575336 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-utilities\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.575916 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcdgw\" (UniqueName: \"kubernetes.io/projected/dc808c2f-935f-4d3f-b282-8b10359f459e-kube-api-access-xcdgw\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.678214 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcdgw\" (UniqueName: \"kubernetes.io/projected/dc808c2f-935f-4d3f-b282-8b10359f459e-kube-api-access-xcdgw\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.678381 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-catalog-content\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.678411 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-utilities\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.678858 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-utilities\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.679368 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-catalog-content\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.699509 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcdgw\" (UniqueName: \"kubernetes.io/projected/dc808c2f-935f-4d3f-b282-8b10359f459e-kube-api-access-xcdgw\") pod \"redhat-marketplace-jvqk6\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:27 crc kubenswrapper[4750]: I0309 18:53:27.845852 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:28 crc kubenswrapper[4750]: I0309 18:53:28.419236 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvqk6"] Mar 09 18:53:29 crc kubenswrapper[4750]: I0309 18:53:29.025245 4750 generic.go:334] "Generic (PLEG): container finished" podID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerID="1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf" exitCode=0 Mar 09 18:53:29 crc kubenswrapper[4750]: I0309 18:53:29.025315 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvqk6" event={"ID":"dc808c2f-935f-4d3f-b282-8b10359f459e","Type":"ContainerDied","Data":"1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf"} Mar 09 18:53:29 crc kubenswrapper[4750]: I0309 18:53:29.025610 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvqk6" event={"ID":"dc808c2f-935f-4d3f-b282-8b10359f459e","Type":"ContainerStarted","Data":"455aa52c084c9a546c2d3384aa7daae4651f4006109f9421664de89fcf4e6d77"} Mar 09 18:53:29 crc kubenswrapper[4750]: I0309 18:53:29.922269 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:30 crc kubenswrapper[4750]: I0309 18:53:30.037825 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvqk6" event={"ID":"dc808c2f-935f-4d3f-b282-8b10359f459e","Type":"ContainerStarted","Data":"824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84"} Mar 09 18:53:31 crc kubenswrapper[4750]: I0309 18:53:31.048096 4750 generic.go:334] "Generic (PLEG): container finished" podID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerID="824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84" exitCode=0 Mar 09 18:53:31 crc kubenswrapper[4750]: I0309 18:53:31.048202 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvqk6" event={"ID":"dc808c2f-935f-4d3f-b282-8b10359f459e","Type":"ContainerDied","Data":"824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84"} Mar 09 18:53:32 crc kubenswrapper[4750]: I0309 18:53:32.061051 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvqk6" event={"ID":"dc808c2f-935f-4d3f-b282-8b10359f459e","Type":"ContainerStarted","Data":"ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736"} Mar 09 18:53:32 crc kubenswrapper[4750]: I0309 18:53:32.094541 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jvqk6" podStartSLOduration=2.63091727 podStartE2EDuration="5.094523459s" podCreationTimestamp="2026-03-09 18:53:27 +0000 UTC" firstStartedPulling="2026-03-09 18:53:29.029088126 +0000 UTC m=+1690.371560524" lastFinishedPulling="2026-03-09 18:53:31.492694325 +0000 UTC m=+1692.835166713" observedRunningTime="2026-03-09 18:53:32.087403285 +0000 UTC m=+1693.429875683" watchObservedRunningTime="2026-03-09 18:53:32.094523459 +0000 UTC m=+1693.436995847" Mar 09 18:53:32 crc kubenswrapper[4750]: I0309 18:53:32.331734 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b6z5n"] Mar 09 18:53:32 crc kubenswrapper[4750]: I0309 18:53:32.332011 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b6z5n" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="registry-server" containerID="cri-o://a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1" gracePeriod=2 Mar 09 18:53:32 crc kubenswrapper[4750]: I0309 18:53:32.840913 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.008988 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-utilities\") pod \"a501006f-283a-4dc1-8498-733c909bdae0\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.009084 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f59jv\" (UniqueName: \"kubernetes.io/projected/a501006f-283a-4dc1-8498-733c909bdae0-kube-api-access-f59jv\") pod \"a501006f-283a-4dc1-8498-733c909bdae0\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.009219 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-catalog-content\") pod \"a501006f-283a-4dc1-8498-733c909bdae0\" (UID: \"a501006f-283a-4dc1-8498-733c909bdae0\") " Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.009929 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-utilities" (OuterVolumeSpecName: "utilities") pod "a501006f-283a-4dc1-8498-733c909bdae0" (UID: "a501006f-283a-4dc1-8498-733c909bdae0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.024781 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a501006f-283a-4dc1-8498-733c909bdae0-kube-api-access-f59jv" (OuterVolumeSpecName: "kube-api-access-f59jv") pod "a501006f-283a-4dc1-8498-733c909bdae0" (UID: "a501006f-283a-4dc1-8498-733c909bdae0"). InnerVolumeSpecName "kube-api-access-f59jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.072343 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a501006f-283a-4dc1-8498-733c909bdae0" (UID: "a501006f-283a-4dc1-8498-733c909bdae0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.079259 4750 generic.go:334] "Generic (PLEG): container finished" podID="a501006f-283a-4dc1-8498-733c909bdae0" containerID="a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1" exitCode=0 Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.079361 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b6z5n" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.079447 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerDied","Data":"a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1"} Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.079491 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b6z5n" event={"ID":"a501006f-283a-4dc1-8498-733c909bdae0","Type":"ContainerDied","Data":"237fcb1e011b481d42164c0ecc0db93990ee262d4d39a1c453fd68922af079ff"} Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.079525 4750 scope.go:117] "RemoveContainer" containerID="a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.122973 4750 scope.go:117] "RemoveContainer" containerID="35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.123975 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.124041 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a501006f-283a-4dc1-8498-733c909bdae0-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.124063 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f59jv\" (UniqueName: \"kubernetes.io/projected/a501006f-283a-4dc1-8498-733c909bdae0-kube-api-access-f59jv\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.145811 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b6z5n"] Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.163548 4750 scope.go:117] "RemoveContainer" containerID="c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.165876 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b6z5n"] Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.204404 4750 scope.go:117] "RemoveContainer" containerID="a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1" Mar 09 18:53:33 crc kubenswrapper[4750]: E0309 18:53:33.205016 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1\": container with ID starting with a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1 not found: ID does not exist" containerID="a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.205068 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1"} err="failed to get container status \"a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1\": rpc error: code = NotFound desc = could not find container \"a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1\": container with ID starting with a05d9c2880f3e937f66f4a41943db140af77bc5fe0b26dcfaca2a8505939ebf1 not found: ID does not exist" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.205105 4750 scope.go:117] "RemoveContainer" containerID="35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10" Mar 09 18:53:33 crc kubenswrapper[4750]: E0309 18:53:33.205459 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10\": container with ID starting with 35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10 not found: ID does not exist" containerID="35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.207051 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10"} err="failed to get container status \"35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10\": rpc error: code = NotFound desc = could not find container \"35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10\": container with ID starting with 35b428cd1c419d9da00bd4755f122fc07c0d7f5066483fcf774856ec16826e10 not found: ID does not exist" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.207133 4750 scope.go:117] "RemoveContainer" containerID="c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07" Mar 09 18:53:33 crc kubenswrapper[4750]: E0309 18:53:33.207549 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07\": container with ID starting with c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07 not found: ID does not exist" containerID="c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.207599 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07"} err="failed to get container status \"c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07\": rpc error: code = NotFound desc = could not find container \"c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07\": container with ID starting with c9a547c6c99e2308cf76b6e08aa9d61e8b97b5dcf0d435b1b3700887611f8c07 not found: ID does not exist" Mar 09 18:53:33 crc kubenswrapper[4750]: I0309 18:53:33.385464 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a501006f-283a-4dc1-8498-733c909bdae0" path="/var/lib/kubelet/pods/a501006f-283a-4dc1-8498-733c909bdae0/volumes" Mar 09 18:53:36 crc kubenswrapper[4750]: I0309 18:53:36.709591 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.338197 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6kdxn"] Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.339113 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6kdxn" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="registry-server" containerID="cri-o://14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534" gracePeriod=2 Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.847353 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.847608 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.898203 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.905370 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.941083 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-catalog-content\") pod \"622ca8d5-72d9-4fce-ad90-9f46ce680820\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.941218 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-utilities\") pod \"622ca8d5-72d9-4fce-ad90-9f46ce680820\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.941276 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwp9p\" (UniqueName: \"kubernetes.io/projected/622ca8d5-72d9-4fce-ad90-9f46ce680820-kube-api-access-cwp9p\") pod \"622ca8d5-72d9-4fce-ad90-9f46ce680820\" (UID: \"622ca8d5-72d9-4fce-ad90-9f46ce680820\") " Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.941845 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-utilities" (OuterVolumeSpecName: "utilities") pod "622ca8d5-72d9-4fce-ad90-9f46ce680820" (UID: "622ca8d5-72d9-4fce-ad90-9f46ce680820"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.958418 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/622ca8d5-72d9-4fce-ad90-9f46ce680820-kube-api-access-cwp9p" (OuterVolumeSpecName: "kube-api-access-cwp9p") pod "622ca8d5-72d9-4fce-ad90-9f46ce680820" (UID: "622ca8d5-72d9-4fce-ad90-9f46ce680820"). InnerVolumeSpecName "kube-api-access-cwp9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:53:37 crc kubenswrapper[4750]: I0309 18:53:37.995038 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "622ca8d5-72d9-4fce-ad90-9f46ce680820" (UID: "622ca8d5-72d9-4fce-ad90-9f46ce680820"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.043175 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.043391 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622ca8d5-72d9-4fce-ad90-9f46ce680820-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.043452 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwp9p\" (UniqueName: \"kubernetes.io/projected/622ca8d5-72d9-4fce-ad90-9f46ce680820-kube-api-access-cwp9p\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.152405 4750 generic.go:334] "Generic (PLEG): container finished" podID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerID="14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534" exitCode=0 Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.152728 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kdxn" event={"ID":"622ca8d5-72d9-4fce-ad90-9f46ce680820","Type":"ContainerDied","Data":"14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534"} Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.152758 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kdxn" event={"ID":"622ca8d5-72d9-4fce-ad90-9f46ce680820","Type":"ContainerDied","Data":"5d62135ec060099c5dff7f3503b0ca4316fe6a03b7c36d93f1834af59e4ae711"} Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.152775 4750 scope.go:117] "RemoveContainer" containerID="14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.153152 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kdxn" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.184744 4750 scope.go:117] "RemoveContainer" containerID="7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.214869 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6kdxn"] Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.216730 4750 scope.go:117] "RemoveContainer" containerID="fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.227780 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6kdxn"] Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.238433 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.302768 4750 scope.go:117] "RemoveContainer" containerID="14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534" Mar 09 18:53:38 crc kubenswrapper[4750]: E0309 18:53:38.303237 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534\": container with ID starting with 14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534 not found: ID does not exist" containerID="14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.303269 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534"} err="failed to get container status \"14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534\": rpc error: code = NotFound desc = could not find container \"14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534\": container with ID starting with 14265d651d45be1c91b26e4de45d1d874ca9521c49769d338aaa2ac837ce4534 not found: ID does not exist" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.303293 4750 scope.go:117] "RemoveContainer" containerID="7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761" Mar 09 18:53:38 crc kubenswrapper[4750]: E0309 18:53:38.303918 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761\": container with ID starting with 7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761 not found: ID does not exist" containerID="7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.303986 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761"} err="failed to get container status \"7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761\": rpc error: code = NotFound desc = could not find container \"7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761\": container with ID starting with 7db822693df44ead36bc7da91e61e9adefcfb2c471bd568666dac5e05ac8a761 not found: ID does not exist" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.304029 4750 scope.go:117] "RemoveContainer" containerID="fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b" Mar 09 18:53:38 crc kubenswrapper[4750]: E0309 18:53:38.304525 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b\": container with ID starting with fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b not found: ID does not exist" containerID="fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b" Mar 09 18:53:38 crc kubenswrapper[4750]: I0309 18:53:38.304562 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b"} err="failed to get container status \"fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b\": rpc error: code = NotFound desc = could not find container \"fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b\": container with ID starting with fc90073ff3a041db3b695e45a9f468dd8643972e0045c6faed46f7e66015f57b not found: ID does not exist" Mar 09 18:53:39 crc kubenswrapper[4750]: I0309 18:53:39.388080 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" path="/var/lib/kubelet/pods/622ca8d5-72d9-4fce-ad90-9f46ce680820/volumes" Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.136346 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvqk6"] Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.173766 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jvqk6" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="registry-server" containerID="cri-o://ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736" gracePeriod=2 Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.724004 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.925432 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcdgw\" (UniqueName: \"kubernetes.io/projected/dc808c2f-935f-4d3f-b282-8b10359f459e-kube-api-access-xcdgw\") pod \"dc808c2f-935f-4d3f-b282-8b10359f459e\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.925738 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-catalog-content\") pod \"dc808c2f-935f-4d3f-b282-8b10359f459e\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.925836 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-utilities\") pod \"dc808c2f-935f-4d3f-b282-8b10359f459e\" (UID: \"dc808c2f-935f-4d3f-b282-8b10359f459e\") " Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.926947 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-utilities" (OuterVolumeSpecName: "utilities") pod "dc808c2f-935f-4d3f-b282-8b10359f459e" (UID: "dc808c2f-935f-4d3f-b282-8b10359f459e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.931885 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc808c2f-935f-4d3f-b282-8b10359f459e-kube-api-access-xcdgw" (OuterVolumeSpecName: "kube-api-access-xcdgw") pod "dc808c2f-935f-4d3f-b282-8b10359f459e" (UID: "dc808c2f-935f-4d3f-b282-8b10359f459e"). InnerVolumeSpecName "kube-api-access-xcdgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:53:40 crc kubenswrapper[4750]: I0309 18:53:40.957183 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc808c2f-935f-4d3f-b282-8b10359f459e" (UID: "dc808c2f-935f-4d3f-b282-8b10359f459e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.028878 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcdgw\" (UniqueName: \"kubernetes.io/projected/dc808c2f-935f-4d3f-b282-8b10359f459e-kube-api-access-xcdgw\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.029172 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.029254 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc808c2f-935f-4d3f-b282-8b10359f459e-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.190326 4750 generic.go:334] "Generic (PLEG): container finished" podID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerID="ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736" exitCode=0 Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.190408 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvqk6" event={"ID":"dc808c2f-935f-4d3f-b282-8b10359f459e","Type":"ContainerDied","Data":"ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736"} Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.190462 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvqk6" event={"ID":"dc808c2f-935f-4d3f-b282-8b10359f459e","Type":"ContainerDied","Data":"455aa52c084c9a546c2d3384aa7daae4651f4006109f9421664de89fcf4e6d77"} Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.190497 4750 scope.go:117] "RemoveContainer" containerID="ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.190585 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvqk6" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.215947 4750 scope.go:117] "RemoveContainer" containerID="824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.251763 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvqk6"] Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.255538 4750 scope.go:117] "RemoveContainer" containerID="1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.261459 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvqk6"] Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.296165 4750 scope.go:117] "RemoveContainer" containerID="ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736" Mar 09 18:53:41 crc kubenswrapper[4750]: E0309 18:53:41.296676 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736\": container with ID starting with ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736 not found: ID does not exist" containerID="ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.296709 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736"} err="failed to get container status \"ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736\": rpc error: code = NotFound desc = could not find container \"ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736\": container with ID starting with ff14dd36efd377fe248a3b71568853f27a016d33d79e4e7f3c00d3169412f736 not found: ID does not exist" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.296734 4750 scope.go:117] "RemoveContainer" containerID="824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84" Mar 09 18:53:41 crc kubenswrapper[4750]: E0309 18:53:41.297122 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84\": container with ID starting with 824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84 not found: ID does not exist" containerID="824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.297144 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84"} err="failed to get container status \"824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84\": rpc error: code = NotFound desc = could not find container \"824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84\": container with ID starting with 824ed134441278fd90a54ff56d107b30e62354cd5aeb5b41e8337c2276a8bb84 not found: ID does not exist" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.297160 4750 scope.go:117] "RemoveContainer" containerID="1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf" Mar 09 18:53:41 crc kubenswrapper[4750]: E0309 18:53:41.297578 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf\": container with ID starting with 1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf not found: ID does not exist" containerID="1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.297606 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf"} err="failed to get container status \"1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf\": rpc error: code = NotFound desc = could not find container \"1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf\": container with ID starting with 1bf0463ded997549c140d8246f67a958cc56768ec193c653a15e142c63f6ecbf not found: ID does not exist" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.386601 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" path="/var/lib/kubelet/pods/dc808c2f-935f-4d3f-b282-8b10359f459e/volumes" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.817767 4750 scope.go:117] "RemoveContainer" containerID="209d4a9e62894dda18feac538f1f3c0dd3d0ff35260741bfb279223d7d96bcbd" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.864337 4750 scope.go:117] "RemoveContainer" containerID="f9d8d985d34814bb2928cc6886169310c947dcea0cf7abd0d8398aa3e3a01f7c" Mar 09 18:53:41 crc kubenswrapper[4750]: I0309 18:53:41.910866 4750 scope.go:117] "RemoveContainer" containerID="5ad8d0972171dc5644b0c875da08f240816d847855e0e2602a5e8575930a3e3e" Mar 09 18:53:51 crc kubenswrapper[4750]: I0309 18:53:51.743816 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:53:51 crc kubenswrapper[4750]: I0309 18:53:51.744460 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.156509 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551374-w4k5b"] Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.158894 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.159014 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.159096 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="extract-utilities" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.159182 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="extract-utilities" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.159269 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="extract-content" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.159352 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="extract-content" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.159447 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="extract-utilities" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.159530 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="extract-utilities" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.159651 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="extract-utilities" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.159754 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="extract-utilities" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.159847 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="extract-content" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.159924 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="extract-content" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.160009 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.160091 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.160187 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.160265 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: E0309 18:54:00.160350 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="extract-content" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.160428 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="extract-content" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.160814 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="622ca8d5-72d9-4fce-ad90-9f46ce680820" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.160929 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a501006f-283a-4dc1-8498-733c909bdae0" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.161019 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc808c2f-935f-4d3f-b282-8b10359f459e" containerName="registry-server" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.162005 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.164769 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.166905 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.167325 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.183265 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551374-w4k5b"] Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.282703 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nct8r\" (UniqueName: \"kubernetes.io/projected/f6200983-07bd-48a6-af4c-2bbfa63afdb2-kube-api-access-nct8r\") pod \"auto-csr-approver-29551374-w4k5b\" (UID: \"f6200983-07bd-48a6-af4c-2bbfa63afdb2\") " pod="openshift-infra/auto-csr-approver-29551374-w4k5b" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.385295 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nct8r\" (UniqueName: \"kubernetes.io/projected/f6200983-07bd-48a6-af4c-2bbfa63afdb2-kube-api-access-nct8r\") pod \"auto-csr-approver-29551374-w4k5b\" (UID: \"f6200983-07bd-48a6-af4c-2bbfa63afdb2\") " pod="openshift-infra/auto-csr-approver-29551374-w4k5b" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.412180 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nct8r\" (UniqueName: \"kubernetes.io/projected/f6200983-07bd-48a6-af4c-2bbfa63afdb2-kube-api-access-nct8r\") pod \"auto-csr-approver-29551374-w4k5b\" (UID: \"f6200983-07bd-48a6-af4c-2bbfa63afdb2\") " pod="openshift-infra/auto-csr-approver-29551374-w4k5b" Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.489874 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" Mar 09 18:54:00 crc kubenswrapper[4750]: W0309 18:54:00.836715 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6200983_07bd_48a6_af4c_2bbfa63afdb2.slice/crio-d8f57e148922b269517e21cc3739c6f86ee991620435e3c20e5325de3fe4c845 WatchSource:0}: Error finding container d8f57e148922b269517e21cc3739c6f86ee991620435e3c20e5325de3fe4c845: Status 404 returned error can't find the container with id d8f57e148922b269517e21cc3739c6f86ee991620435e3c20e5325de3fe4c845 Mar 09 18:54:00 crc kubenswrapper[4750]: I0309 18:54:00.838577 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551374-w4k5b"] Mar 09 18:54:01 crc kubenswrapper[4750]: I0309 18:54:01.455728 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" event={"ID":"f6200983-07bd-48a6-af4c-2bbfa63afdb2","Type":"ContainerStarted","Data":"d8f57e148922b269517e21cc3739c6f86ee991620435e3c20e5325de3fe4c845"} Mar 09 18:54:02 crc kubenswrapper[4750]: I0309 18:54:02.466871 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" event={"ID":"f6200983-07bd-48a6-af4c-2bbfa63afdb2","Type":"ContainerStarted","Data":"99e4ea5b5a58c028cf87f4fd08f6ab66a9f2db77da502f4d2b163a9c1538efad"} Mar 09 18:54:02 crc kubenswrapper[4750]: I0309 18:54:02.486914 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" podStartSLOduration=1.304797997 podStartE2EDuration="2.486893088s" podCreationTimestamp="2026-03-09 18:54:00 +0000 UTC" firstStartedPulling="2026-03-09 18:54:00.839944409 +0000 UTC m=+1722.182416817" lastFinishedPulling="2026-03-09 18:54:02.0220395 +0000 UTC m=+1723.364511908" observedRunningTime="2026-03-09 18:54:02.479871087 +0000 UTC m=+1723.822343485" watchObservedRunningTime="2026-03-09 18:54:02.486893088 +0000 UTC m=+1723.829365486" Mar 09 18:54:03 crc kubenswrapper[4750]: I0309 18:54:03.486994 4750 generic.go:334] "Generic (PLEG): container finished" podID="f6200983-07bd-48a6-af4c-2bbfa63afdb2" containerID="99e4ea5b5a58c028cf87f4fd08f6ab66a9f2db77da502f4d2b163a9c1538efad" exitCode=0 Mar 09 18:54:03 crc kubenswrapper[4750]: I0309 18:54:03.487080 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" event={"ID":"f6200983-07bd-48a6-af4c-2bbfa63afdb2","Type":"ContainerDied","Data":"99e4ea5b5a58c028cf87f4fd08f6ab66a9f2db77da502f4d2b163a9c1538efad"} Mar 09 18:54:04 crc kubenswrapper[4750]: I0309 18:54:04.928586 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.090692 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nct8r\" (UniqueName: \"kubernetes.io/projected/f6200983-07bd-48a6-af4c-2bbfa63afdb2-kube-api-access-nct8r\") pod \"f6200983-07bd-48a6-af4c-2bbfa63afdb2\" (UID: \"f6200983-07bd-48a6-af4c-2bbfa63afdb2\") " Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.099411 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6200983-07bd-48a6-af4c-2bbfa63afdb2-kube-api-access-nct8r" (OuterVolumeSpecName: "kube-api-access-nct8r") pod "f6200983-07bd-48a6-af4c-2bbfa63afdb2" (UID: "f6200983-07bd-48a6-af4c-2bbfa63afdb2"). InnerVolumeSpecName "kube-api-access-nct8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.194346 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nct8r\" (UniqueName: \"kubernetes.io/projected/f6200983-07bd-48a6-af4c-2bbfa63afdb2-kube-api-access-nct8r\") on node \"crc\" DevicePath \"\"" Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.515808 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" event={"ID":"f6200983-07bd-48a6-af4c-2bbfa63afdb2","Type":"ContainerDied","Data":"d8f57e148922b269517e21cc3739c6f86ee991620435e3c20e5325de3fe4c845"} Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.515871 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8f57e148922b269517e21cc3739c6f86ee991620435e3c20e5325de3fe4c845" Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.515966 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551374-w4k5b" Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.582566 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551368-ngvhp"] Mar 09 18:54:05 crc kubenswrapper[4750]: I0309 18:54:05.596378 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551368-ngvhp"] Mar 09 18:54:07 crc kubenswrapper[4750]: I0309 18:54:07.385482 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d81c74f1-ee19-4573-8582-e2fdf6373472" path="/var/lib/kubelet/pods/d81c74f1-ee19-4573-8582-e2fdf6373472/volumes" Mar 09 18:54:21 crc kubenswrapper[4750]: I0309 18:54:21.743729 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 18:54:21 crc kubenswrapper[4750]: I0309 18:54:21.744825 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 18:54:21 crc kubenswrapper[4750]: I0309 18:54:21.744904 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 18:54:21 crc kubenswrapper[4750]: I0309 18:54:21.746215 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 18:54:21 crc kubenswrapper[4750]: I0309 18:54:21.746304 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" gracePeriod=600 Mar 09 18:54:21 crc kubenswrapper[4750]: E0309 18:54:21.907680 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:54:22 crc kubenswrapper[4750]: I0309 18:54:22.738069 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" exitCode=0 Mar 09 18:54:22 crc kubenswrapper[4750]: I0309 18:54:22.738177 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1"} Mar 09 18:54:22 crc kubenswrapper[4750]: I0309 18:54:22.738479 4750 scope.go:117] "RemoveContainer" containerID="0bf642275aba662d4b9920780da985536f4c7bcb2a54a88c1149087655cecbbd" Mar 09 18:54:22 crc kubenswrapper[4750]: I0309 18:54:22.739596 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:54:22 crc kubenswrapper[4750]: E0309 18:54:22.740068 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:54:38 crc kubenswrapper[4750]: I0309 18:54:38.374540 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:54:38 crc kubenswrapper[4750]: E0309 18:54:38.375596 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:54:42 crc kubenswrapper[4750]: I0309 18:54:42.065501 4750 scope.go:117] "RemoveContainer" containerID="846929547db069fe8adbbcf0a02d640976c07c15d2e9d7eb9ef85fd81f1a0aa7" Mar 09 18:54:42 crc kubenswrapper[4750]: I0309 18:54:42.108841 4750 scope.go:117] "RemoveContainer" containerID="b2aef219fc788d84813540d7b11774c8bf24d266e97ec717a2d9c1cee4a0e09e" Mar 09 18:54:42 crc kubenswrapper[4750]: I0309 18:54:42.143981 4750 scope.go:117] "RemoveContainer" containerID="85ec18044d65e2b9a602eb3c76782ebc70dd51b5f4e6463deea04d551c4f57a3" Mar 09 18:54:51 crc kubenswrapper[4750]: I0309 18:54:51.373716 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:54:51 crc kubenswrapper[4750]: E0309 18:54:51.374703 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:55:03 crc kubenswrapper[4750]: I0309 18:55:03.374151 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:55:03 crc kubenswrapper[4750]: E0309 18:55:03.375001 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:55:18 crc kubenswrapper[4750]: I0309 18:55:18.373975 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:55:18 crc kubenswrapper[4750]: E0309 18:55:18.374799 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:55:31 crc kubenswrapper[4750]: I0309 18:55:31.374107 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:55:31 crc kubenswrapper[4750]: E0309 18:55:31.375096 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:55:44 crc kubenswrapper[4750]: I0309 18:55:44.374076 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:55:44 crc kubenswrapper[4750]: E0309 18:55:44.375365 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:55:56 crc kubenswrapper[4750]: I0309 18:55:56.373951 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:55:56 crc kubenswrapper[4750]: E0309 18:55:56.375231 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.190454 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551376-ckf8j"] Mar 09 18:56:00 crc kubenswrapper[4750]: E0309 18:56:00.192567 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6200983-07bd-48a6-af4c-2bbfa63afdb2" containerName="oc" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.192598 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6200983-07bd-48a6-af4c-2bbfa63afdb2" containerName="oc" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.193890 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6200983-07bd-48a6-af4c-2bbfa63afdb2" containerName="oc" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.201675 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.206995 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.207171 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.207558 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.216583 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551376-ckf8j"] Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.276661 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gtlh\" (UniqueName: \"kubernetes.io/projected/d826dd46-ddf0-4c54-a075-f59baae9aa35-kube-api-access-6gtlh\") pod \"auto-csr-approver-29551376-ckf8j\" (UID: \"d826dd46-ddf0-4c54-a075-f59baae9aa35\") " pod="openshift-infra/auto-csr-approver-29551376-ckf8j" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.378669 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gtlh\" (UniqueName: \"kubernetes.io/projected/d826dd46-ddf0-4c54-a075-f59baae9aa35-kube-api-access-6gtlh\") pod \"auto-csr-approver-29551376-ckf8j\" (UID: \"d826dd46-ddf0-4c54-a075-f59baae9aa35\") " pod="openshift-infra/auto-csr-approver-29551376-ckf8j" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.405952 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gtlh\" (UniqueName: \"kubernetes.io/projected/d826dd46-ddf0-4c54-a075-f59baae9aa35-kube-api-access-6gtlh\") pod \"auto-csr-approver-29551376-ckf8j\" (UID: \"d826dd46-ddf0-4c54-a075-f59baae9aa35\") " pod="openshift-infra/auto-csr-approver-29551376-ckf8j" Mar 09 18:56:00 crc kubenswrapper[4750]: I0309 18:56:00.537011 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" Mar 09 18:56:01 crc kubenswrapper[4750]: I0309 18:56:01.051904 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551376-ckf8j"] Mar 09 18:56:01 crc kubenswrapper[4750]: I0309 18:56:01.242306 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" event={"ID":"d826dd46-ddf0-4c54-a075-f59baae9aa35","Type":"ContainerStarted","Data":"61a9ea7639b87e801095cfc5f32209331e1c42fcbac4c7ab7168513a5b671e0c"} Mar 09 18:56:03 crc kubenswrapper[4750]: I0309 18:56:03.269342 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" event={"ID":"d826dd46-ddf0-4c54-a075-f59baae9aa35","Type":"ContainerStarted","Data":"53c05d265185f8b21273e90b0774de4a8db394cf52babbb8137ebbeaeffb1b1b"} Mar 09 18:56:03 crc kubenswrapper[4750]: I0309 18:56:03.292265 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" podStartSLOduration=1.496449991 podStartE2EDuration="3.292247186s" podCreationTimestamp="2026-03-09 18:56:00 +0000 UTC" firstStartedPulling="2026-03-09 18:56:01.068904092 +0000 UTC m=+1842.411376490" lastFinishedPulling="2026-03-09 18:56:02.864701297 +0000 UTC m=+1844.207173685" observedRunningTime="2026-03-09 18:56:03.287669521 +0000 UTC m=+1844.630141929" watchObservedRunningTime="2026-03-09 18:56:03.292247186 +0000 UTC m=+1844.634719574" Mar 09 18:56:04 crc kubenswrapper[4750]: I0309 18:56:04.293357 4750 generic.go:334] "Generic (PLEG): container finished" podID="d826dd46-ddf0-4c54-a075-f59baae9aa35" containerID="53c05d265185f8b21273e90b0774de4a8db394cf52babbb8137ebbeaeffb1b1b" exitCode=0 Mar 09 18:56:04 crc kubenswrapper[4750]: I0309 18:56:04.293501 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" event={"ID":"d826dd46-ddf0-4c54-a075-f59baae9aa35","Type":"ContainerDied","Data":"53c05d265185f8b21273e90b0774de4a8db394cf52babbb8137ebbeaeffb1b1b"} Mar 09 18:56:05 crc kubenswrapper[4750]: I0309 18:56:05.699900 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" Mar 09 18:56:05 crc kubenswrapper[4750]: I0309 18:56:05.815366 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gtlh\" (UniqueName: \"kubernetes.io/projected/d826dd46-ddf0-4c54-a075-f59baae9aa35-kube-api-access-6gtlh\") pod \"d826dd46-ddf0-4c54-a075-f59baae9aa35\" (UID: \"d826dd46-ddf0-4c54-a075-f59baae9aa35\") " Mar 09 18:56:05 crc kubenswrapper[4750]: I0309 18:56:05.820486 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d826dd46-ddf0-4c54-a075-f59baae9aa35-kube-api-access-6gtlh" (OuterVolumeSpecName: "kube-api-access-6gtlh") pod "d826dd46-ddf0-4c54-a075-f59baae9aa35" (UID: "d826dd46-ddf0-4c54-a075-f59baae9aa35"). InnerVolumeSpecName "kube-api-access-6gtlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:56:05 crc kubenswrapper[4750]: I0309 18:56:05.918346 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gtlh\" (UniqueName: \"kubernetes.io/projected/d826dd46-ddf0-4c54-a075-f59baae9aa35-kube-api-access-6gtlh\") on node \"crc\" DevicePath \"\"" Mar 09 18:56:06 crc kubenswrapper[4750]: I0309 18:56:06.323386 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" event={"ID":"d826dd46-ddf0-4c54-a075-f59baae9aa35","Type":"ContainerDied","Data":"61a9ea7639b87e801095cfc5f32209331e1c42fcbac4c7ab7168513a5b671e0c"} Mar 09 18:56:06 crc kubenswrapper[4750]: I0309 18:56:06.323842 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61a9ea7639b87e801095cfc5f32209331e1c42fcbac4c7ab7168513a5b671e0c" Mar 09 18:56:06 crc kubenswrapper[4750]: I0309 18:56:06.323478 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551376-ckf8j" Mar 09 18:56:06 crc kubenswrapper[4750]: I0309 18:56:06.385019 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551370-bc4wr"] Mar 09 18:56:06 crc kubenswrapper[4750]: I0309 18:56:06.394275 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551370-bc4wr"] Mar 09 18:56:07 crc kubenswrapper[4750]: I0309 18:56:07.374343 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:56:07 crc kubenswrapper[4750]: E0309 18:56:07.375109 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:56:07 crc kubenswrapper[4750]: I0309 18:56:07.387598 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38bfb493-9d48-470c-b08f-b792cad68e98" path="/var/lib/kubelet/pods/38bfb493-9d48-470c-b08f-b792cad68e98/volumes" Mar 09 18:56:16 crc kubenswrapper[4750]: I0309 18:56:16.455827 4750 generic.go:334] "Generic (PLEG): container finished" podID="edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" containerID="8b87065ef1e47a9957f2ddc4056e605f9227718859027be5b5520c09f2fa02a0" exitCode=0 Mar 09 18:56:16 crc kubenswrapper[4750]: I0309 18:56:16.455884 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" event={"ID":"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee","Type":"ContainerDied","Data":"8b87065ef1e47a9957f2ddc4056e605f9227718859027be5b5520c09f2fa02a0"} Mar 09 18:56:17 crc kubenswrapper[4750]: I0309 18:56:17.974620 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.124723 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-bootstrap-combined-ca-bundle\") pod \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.124938 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-inventory\") pod \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.125083 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-ssh-key-openstack-edpm-ipam\") pod \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.126140 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t495x\" (UniqueName: \"kubernetes.io/projected/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-kube-api-access-t495x\") pod \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\" (UID: \"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee\") " Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.131519 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-kube-api-access-t495x" (OuterVolumeSpecName: "kube-api-access-t495x") pod "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" (UID: "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee"). InnerVolumeSpecName "kube-api-access-t495x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.132411 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" (UID: "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.156828 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-inventory" (OuterVolumeSpecName: "inventory") pod "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" (UID: "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.177952 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" (UID: "edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.228369 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t495x\" (UniqueName: \"kubernetes.io/projected/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-kube-api-access-t495x\") on node \"crc\" DevicePath \"\"" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.228401 4750 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.228427 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.228438 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.374148 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:56:18 crc kubenswrapper[4750]: E0309 18:56:18.374700 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.500709 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" event={"ID":"edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee","Type":"ContainerDied","Data":"3bb2df63f15b16956ab97e75979433db8aaab92aca16183a0ffbd0de4f305f9a"} Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.500765 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bb2df63f15b16956ab97e75979433db8aaab92aca16183a0ffbd0de4f305f9a" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.500833 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.587366 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm"] Mar 09 18:56:18 crc kubenswrapper[4750]: E0309 18:56:18.588195 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d826dd46-ddf0-4c54-a075-f59baae9aa35" containerName="oc" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.588222 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d826dd46-ddf0-4c54-a075-f59baae9aa35" containerName="oc" Mar 09 18:56:18 crc kubenswrapper[4750]: E0309 18:56:18.588261 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.588270 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.588517 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.588541 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d826dd46-ddf0-4c54-a075-f59baae9aa35" containerName="oc" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.589743 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.592010 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.592014 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.592684 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.596003 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.603306 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm"] Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.742087 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbgmg\" (UniqueName: \"kubernetes.io/projected/825b5a72-cb74-421a-8ca9-103d99691830-kube-api-access-bbgmg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.742342 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.742773 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.845725 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.845954 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbgmg\" (UniqueName: \"kubernetes.io/projected/825b5a72-cb74-421a-8ca9-103d99691830-kube-api-access-bbgmg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.846328 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.852050 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.852470 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.867784 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbgmg\" (UniqueName: \"kubernetes.io/projected/825b5a72-cb74-421a-8ca9-103d99691830-kube-api-access-bbgmg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:18 crc kubenswrapper[4750]: I0309 18:56:18.915205 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:56:19 crc kubenswrapper[4750]: I0309 18:56:19.036260 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1af5-account-create-update-c74fv"] Mar 09 18:56:19 crc kubenswrapper[4750]: I0309 18:56:19.046143 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1af5-account-create-update-c74fv"] Mar 09 18:56:19 crc kubenswrapper[4750]: I0309 18:56:19.397011 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567fa16f-93f1-430f-a041-aba3aaec9957" path="/var/lib/kubelet/pods/567fa16f-93f1-430f-a041-aba3aaec9957/volumes" Mar 09 18:56:19 crc kubenswrapper[4750]: I0309 18:56:19.485671 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm"] Mar 09 18:56:19 crc kubenswrapper[4750]: I0309 18:56:19.528195 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" event={"ID":"825b5a72-cb74-421a-8ca9-103d99691830","Type":"ContainerStarted","Data":"aee5dfd68de2fedcff482e7cbbfbdcf89ca5e402f4f490b7644a5fa59ea108ec"} Mar 09 18:56:19 crc kubenswrapper[4750]: I0309 18:56:19.942511 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:56:20 crc kubenswrapper[4750]: I0309 18:56:20.042720 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-p5r85"] Mar 09 18:56:20 crc kubenswrapper[4750]: I0309 18:56:20.062156 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-p5r85"] Mar 09 18:56:20 crc kubenswrapper[4750]: I0309 18:56:20.074573 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-f9d9-account-create-update-868dt"] Mar 09 18:56:20 crc kubenswrapper[4750]: I0309 18:56:20.085118 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-f9d9-account-create-update-868dt"] Mar 09 18:56:20 crc kubenswrapper[4750]: I0309 18:56:20.543713 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" event={"ID":"825b5a72-cb74-421a-8ca9-103d99691830","Type":"ContainerStarted","Data":"9dcfac7a272df3dd60a516c43132f16c1c44807ea940892dc84e8a2a2ac6a8c9"} Mar 09 18:56:20 crc kubenswrapper[4750]: I0309 18:56:20.570249 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" podStartSLOduration=2.131972524 podStartE2EDuration="2.570228884s" podCreationTimestamp="2026-03-09 18:56:18 +0000 UTC" firstStartedPulling="2026-03-09 18:56:19.500448982 +0000 UTC m=+1860.842921380" lastFinishedPulling="2026-03-09 18:56:19.938705342 +0000 UTC m=+1861.281177740" observedRunningTime="2026-03-09 18:56:20.568826455 +0000 UTC m=+1861.911298903" watchObservedRunningTime="2026-03-09 18:56:20.570228884 +0000 UTC m=+1861.912701282" Mar 09 18:56:21 crc kubenswrapper[4750]: I0309 18:56:21.037056 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-qlj7j"] Mar 09 18:56:21 crc kubenswrapper[4750]: I0309 18:56:21.051039 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-qlj7j"] Mar 09 18:56:21 crc kubenswrapper[4750]: I0309 18:56:21.389731 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c0825fd-c53a-46cc-9b8a-fca000620d76" path="/var/lib/kubelet/pods/1c0825fd-c53a-46cc-9b8a-fca000620d76/volumes" Mar 09 18:56:21 crc kubenswrapper[4750]: I0309 18:56:21.391089 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bf233bd-79f7-4a07-8328-e7491ebb967b" path="/var/lib/kubelet/pods/5bf233bd-79f7-4a07-8328-e7491ebb967b/volumes" Mar 09 18:56:21 crc kubenswrapper[4750]: I0309 18:56:21.392209 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4601d20-845b-4ab0-9f53-0e10aa1b46a7" path="/var/lib/kubelet/pods/f4601d20-845b-4ab0-9f53-0e10aa1b46a7/volumes" Mar 09 18:56:23 crc kubenswrapper[4750]: I0309 18:56:23.043148 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-d55d-account-create-update-rx5b2"] Mar 09 18:56:23 crc kubenswrapper[4750]: I0309 18:56:23.056267 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-lr9tk"] Mar 09 18:56:23 crc kubenswrapper[4750]: I0309 18:56:23.064647 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-d55d-account-create-update-rx5b2"] Mar 09 18:56:23 crc kubenswrapper[4750]: I0309 18:56:23.072882 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-lr9tk"] Mar 09 18:56:23 crc kubenswrapper[4750]: I0309 18:56:23.387512 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7" path="/var/lib/kubelet/pods/9ce1e27e-5b44-4474-9f57-e6eaa7f75cf7/volumes" Mar 09 18:56:23 crc kubenswrapper[4750]: I0309 18:56:23.389167 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaedb9df-9a83-4ef4-bb62-8fac20a5cb20" path="/var/lib/kubelet/pods/eaedb9df-9a83-4ef4-bb62-8fac20a5cb20/volumes" Mar 09 18:56:30 crc kubenswrapper[4750]: I0309 18:56:30.037404 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-10ba-account-create-update-5p6zt"] Mar 09 18:56:30 crc kubenswrapper[4750]: I0309 18:56:30.051326 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-hnrfj"] Mar 09 18:56:30 crc kubenswrapper[4750]: I0309 18:56:30.065278 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-10ba-account-create-update-5p6zt"] Mar 09 18:56:30 crc kubenswrapper[4750]: I0309 18:56:30.080257 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-hnrfj"] Mar 09 18:56:31 crc kubenswrapper[4750]: I0309 18:56:31.388386 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33e86472-4600-45ec-888c-988c15731c69" path="/var/lib/kubelet/pods/33e86472-4600-45ec-888c-988c15731c69/volumes" Mar 09 18:56:31 crc kubenswrapper[4750]: I0309 18:56:31.389942 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f192f0ed-c217-472f-9535-91518a9337d4" path="/var/lib/kubelet/pods/f192f0ed-c217-472f-9535-91518a9337d4/volumes" Mar 09 18:56:33 crc kubenswrapper[4750]: I0309 18:56:33.373803 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:56:33 crc kubenswrapper[4750]: E0309 18:56:33.374589 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.292470 4750 scope.go:117] "RemoveContainer" containerID="fa5afb0335b88ea964b7f114e452dfd93c4085a1cf39e4d92754ed47642d3ccd" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.332824 4750 scope.go:117] "RemoveContainer" containerID="38d9b0b593126dc4246880798f9c506666e03f01b7e244012505849cadf74b0a" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.403559 4750 scope.go:117] "RemoveContainer" containerID="3477901d8ff8b9596d5a273aeaf7f08f34e9f6ef08b3a89e9699fd44163c6ce2" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.463241 4750 scope.go:117] "RemoveContainer" containerID="0464bab5fa574cfb5bba01a88ed142ed4f2c024f773530f5d5b22f79d500ba13" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.530287 4750 scope.go:117] "RemoveContainer" containerID="e52a5353a572e957bfb271bd766b9b481f27132b075183bed227dcef209ef206" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.558937 4750 scope.go:117] "RemoveContainer" containerID="af8da8c525aa15597d26252e499384ac9131562117fee699016088d024d65f34" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.597368 4750 scope.go:117] "RemoveContainer" containerID="0a0acb2e286349de69f28fd4e370cbbf4a675cc106e8ee016b6370abdaeac43c" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.630304 4750 scope.go:117] "RemoveContainer" containerID="5bf3cee0e9565617eecc6eb8fdf568363527ad515dd9eab1ccdab747319a3e42" Mar 09 18:56:42 crc kubenswrapper[4750]: I0309 18:56:42.656147 4750 scope.go:117] "RemoveContainer" containerID="f144559cfb61a2b80f797c7a32494d41b8d4cc0ce2b5220a351212b31cef59bf" Mar 09 18:56:44 crc kubenswrapper[4750]: I0309 18:56:44.047677 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-mqfgs"] Mar 09 18:56:44 crc kubenswrapper[4750]: I0309 18:56:44.060511 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-mqfgs"] Mar 09 18:56:44 crc kubenswrapper[4750]: I0309 18:56:44.372889 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:56:44 crc kubenswrapper[4750]: E0309 18:56:44.373263 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:56:45 crc kubenswrapper[4750]: I0309 18:56:45.385973 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="783ff672-9bd9-4092-8823-9a7bc9d2364a" path="/var/lib/kubelet/pods/783ff672-9bd9-4092-8823-9a7bc9d2364a/volumes" Mar 09 18:56:46 crc kubenswrapper[4750]: I0309 18:56:46.049666 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-ba6b-account-create-update-z9h5p"] Mar 09 18:56:46 crc kubenswrapper[4750]: I0309 18:56:46.066450 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-ba6b-account-create-update-z9h5p"] Mar 09 18:56:47 crc kubenswrapper[4750]: I0309 18:56:47.042580 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-mzvtm"] Mar 09 18:56:47 crc kubenswrapper[4750]: I0309 18:56:47.054126 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rfqnd"] Mar 09 18:56:47 crc kubenswrapper[4750]: I0309 18:56:47.066597 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rfqnd"] Mar 09 18:56:47 crc kubenswrapper[4750]: I0309 18:56:47.076130 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-mzvtm"] Mar 09 18:56:47 crc kubenswrapper[4750]: I0309 18:56:47.398135 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6" path="/var/lib/kubelet/pods/3dfc9777-b33a-4f5b-bfc0-c7c7370e37e6/volumes" Mar 09 18:56:47 crc kubenswrapper[4750]: I0309 18:56:47.399044 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59de480c-59b5-4a00-92c1-88c8fbc79366" path="/var/lib/kubelet/pods/59de480c-59b5-4a00-92c1-88c8fbc79366/volumes" Mar 09 18:56:47 crc kubenswrapper[4750]: I0309 18:56:47.399975 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c94fb97-754d-494d-ac4a-9fc9f3c9022c" path="/var/lib/kubelet/pods/6c94fb97-754d-494d-ac4a-9fc9f3c9022c/volumes" Mar 09 18:56:51 crc kubenswrapper[4750]: I0309 18:56:51.039402 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-57gqs"] Mar 09 18:56:51 crc kubenswrapper[4750]: I0309 18:56:51.051398 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-57gqs"] Mar 09 18:56:51 crc kubenswrapper[4750]: I0309 18:56:51.392470 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6449a6c0-9e32-4b3c-b513-9672b6edb485" path="/var/lib/kubelet/pods/6449a6c0-9e32-4b3c-b513-9672b6edb485/volumes" Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.060863 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78c4-account-create-update-8v6f7"] Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.078921 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8k6sd"] Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.090352 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c547-account-create-update-4tbqv"] Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.098596 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8k6sd"] Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.106333 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-78c4-account-create-update-8v6f7"] Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.115037 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c547-account-create-update-4tbqv"] Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.396603 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e796a39-dd3b-4a32-ab3b-3c9e341381fb" path="/var/lib/kubelet/pods/0e796a39-dd3b-4a32-ab3b-3c9e341381fb/volumes" Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.397862 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46354cf1-acc8-45a7-a1e7-92f65230b565" path="/var/lib/kubelet/pods/46354cf1-acc8-45a7-a1e7-92f65230b565/volumes" Mar 09 18:56:55 crc kubenswrapper[4750]: I0309 18:56:55.399278 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f69cd0cb-75f0-4e75-825b-3f578b98f056" path="/var/lib/kubelet/pods/f69cd0cb-75f0-4e75-825b-3f578b98f056/volumes" Mar 09 18:56:57 crc kubenswrapper[4750]: I0309 18:56:57.374608 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:56:57 crc kubenswrapper[4750]: E0309 18:56:57.374936 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:57:02 crc kubenswrapper[4750]: I0309 18:57:02.053289 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-m55h9"] Mar 09 18:57:02 crc kubenswrapper[4750]: I0309 18:57:02.066407 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-m55h9"] Mar 09 18:57:02 crc kubenswrapper[4750]: I0309 18:57:02.082695 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ln7q6"] Mar 09 18:57:02 crc kubenswrapper[4750]: I0309 18:57:02.090739 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ln7q6"] Mar 09 18:57:03 crc kubenswrapper[4750]: I0309 18:57:03.393192 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a84f925-3ae7-44b6-9e09-76da4848f588" path="/var/lib/kubelet/pods/8a84f925-3ae7-44b6-9e09-76da4848f588/volumes" Mar 09 18:57:03 crc kubenswrapper[4750]: I0309 18:57:03.395105 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d34acaa2-45bf-476e-8727-27b788e2887e" path="/var/lib/kubelet/pods/d34acaa2-45bf-476e-8727-27b788e2887e/volumes" Mar 09 18:57:09 crc kubenswrapper[4750]: I0309 18:57:09.374425 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:57:09 crc kubenswrapper[4750]: E0309 18:57:09.375615 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:57:21 crc kubenswrapper[4750]: I0309 18:57:21.374795 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:57:21 crc kubenswrapper[4750]: E0309 18:57:21.375726 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:57:36 crc kubenswrapper[4750]: I0309 18:57:36.374380 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:57:36 crc kubenswrapper[4750]: E0309 18:57:36.375390 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:57:37 crc kubenswrapper[4750]: I0309 18:57:37.053719 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-x6m2l"] Mar 09 18:57:37 crc kubenswrapper[4750]: I0309 18:57:37.068620 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-x6m2l"] Mar 09 18:57:37 crc kubenswrapper[4750]: I0309 18:57:37.384670 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="620fb043-252f-4bdc-8120-68a3978e1c82" path="/var/lib/kubelet/pods/620fb043-252f-4bdc-8120-68a3978e1c82/volumes" Mar 09 18:57:42 crc kubenswrapper[4750]: I0309 18:57:42.897485 4750 scope.go:117] "RemoveContainer" containerID="bf9e1f72b2a0fcea54ff0ed44e2f2dc6919d164030ad2da3a7deb912556e0faa" Mar 09 18:57:42 crc kubenswrapper[4750]: I0309 18:57:42.945080 4750 scope.go:117] "RemoveContainer" containerID="1f493884b398fc5c6bbca68a79dd5d8d92588c11220019ac57895b7872c5be67" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.024132 4750 scope.go:117] "RemoveContainer" containerID="6abb1c45bded4a1262afdeac6a5fe08053681bf937d7ae14cdcbfa48e057c5f9" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.074321 4750 scope.go:117] "RemoveContainer" containerID="f73e1c4aa1476494bae0883ca8680b278dc6dc9f17f0cb23e2ecc770967f6191" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.117866 4750 scope.go:117] "RemoveContainer" containerID="79d793d41819e39872483071d0df21fac11e13760912f8b4ee73535605ed9d8a" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.194164 4750 scope.go:117] "RemoveContainer" containerID="c1c097139ef200f5474967dd9000e7eb1664f09c25fc16f9adfbce7a95aae746" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.256708 4750 scope.go:117] "RemoveContainer" containerID="43104fa1b64d7ee48e967155b623a9515efab604fcddbaddbd8f401c10b0bf98" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.301049 4750 scope.go:117] "RemoveContainer" containerID="90277251ba6f9ada77e72495f221957c4824f976e512b472ce77961b9540d532" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.328054 4750 scope.go:117] "RemoveContainer" containerID="0fb3a6126fd1cc367974868591fa5353b289803c04067a47eb4f58f16eb88a99" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.368293 4750 scope.go:117] "RemoveContainer" containerID="0c84c96140e622b0b2db5d836b37df651d1396b87c87bdee6df1263672f98fde" Mar 09 18:57:43 crc kubenswrapper[4750]: I0309 18:57:43.399397 4750 scope.go:117] "RemoveContainer" containerID="1dff630734ab0900c47fda83ca71c84f6f2fee99fc3c331dc1ce4956c779f4e1" Mar 09 18:57:48 crc kubenswrapper[4750]: I0309 18:57:48.690559 4750 generic.go:334] "Generic (PLEG): container finished" podID="825b5a72-cb74-421a-8ca9-103d99691830" containerID="9dcfac7a272df3dd60a516c43132f16c1c44807ea940892dc84e8a2a2ac6a8c9" exitCode=0 Mar 09 18:57:48 crc kubenswrapper[4750]: I0309 18:57:48.690755 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" event={"ID":"825b5a72-cb74-421a-8ca9-103d99691830","Type":"ContainerDied","Data":"9dcfac7a272df3dd60a516c43132f16c1c44807ea940892dc84e8a2a2ac6a8c9"} Mar 09 18:57:49 crc kubenswrapper[4750]: I0309 18:57:49.384161 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:57:49 crc kubenswrapper[4750]: E0309 18:57:49.385022 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.047432 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hsgm7"] Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.056283 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hsgm7"] Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.083822 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-59wvm"] Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.091978 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-59wvm"] Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.100456 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t8lkt"] Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.107393 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t8lkt"] Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.167019 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.201961 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-inventory\") pod \"825b5a72-cb74-421a-8ca9-103d99691830\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.202066 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-ssh-key-openstack-edpm-ipam\") pod \"825b5a72-cb74-421a-8ca9-103d99691830\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.202116 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbgmg\" (UniqueName: \"kubernetes.io/projected/825b5a72-cb74-421a-8ca9-103d99691830-kube-api-access-bbgmg\") pod \"825b5a72-cb74-421a-8ca9-103d99691830\" (UID: \"825b5a72-cb74-421a-8ca9-103d99691830\") " Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.208865 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825b5a72-cb74-421a-8ca9-103d99691830-kube-api-access-bbgmg" (OuterVolumeSpecName: "kube-api-access-bbgmg") pod "825b5a72-cb74-421a-8ca9-103d99691830" (UID: "825b5a72-cb74-421a-8ca9-103d99691830"). InnerVolumeSpecName "kube-api-access-bbgmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.230467 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-inventory" (OuterVolumeSpecName: "inventory") pod "825b5a72-cb74-421a-8ca9-103d99691830" (UID: "825b5a72-cb74-421a-8ca9-103d99691830"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.236377 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "825b5a72-cb74-421a-8ca9-103d99691830" (UID: "825b5a72-cb74-421a-8ca9-103d99691830"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.304833 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.304867 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbgmg\" (UniqueName: \"kubernetes.io/projected/825b5a72-cb74-421a-8ca9-103d99691830-kube-api-access-bbgmg\") on node \"crc\" DevicePath \"\"" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.304876 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825b5a72-cb74-421a-8ca9-103d99691830-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.721434 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" event={"ID":"825b5a72-cb74-421a-8ca9-103d99691830","Type":"ContainerDied","Data":"aee5dfd68de2fedcff482e7cbbfbdcf89ca5e402f4f490b7644a5fa59ea108ec"} Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.721474 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.721499 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aee5dfd68de2fedcff482e7cbbfbdcf89ca5e402f4f490b7644a5fa59ea108ec" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.834083 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p"] Mar 09 18:57:50 crc kubenswrapper[4750]: E0309 18:57:50.835337 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825b5a72-cb74-421a-8ca9-103d99691830" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.835382 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="825b5a72-cb74-421a-8ca9-103d99691830" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.836009 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="825b5a72-cb74-421a-8ca9-103d99691830" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.837026 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.840320 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.840560 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.840673 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.843468 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.859005 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p"] Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.934566 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.934660 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77jjv\" (UniqueName: \"kubernetes.io/projected/ea9561e1-8a07-4d8a-9726-d1e06b93af82-kube-api-access-77jjv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:50 crc kubenswrapper[4750]: I0309 18:57:50.934719 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.036712 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.037016 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77jjv\" (UniqueName: \"kubernetes.io/projected/ea9561e1-8a07-4d8a-9726-d1e06b93af82-kube-api-access-77jjv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.037182 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.040923 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.041257 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.053686 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77jjv\" (UniqueName: \"kubernetes.io/projected/ea9561e1-8a07-4d8a-9726-d1e06b93af82-kube-api-access-77jjv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7t27p\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.153730 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.392768 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="215dfd5c-6d06-42ef-89c7-6d40910fbf28" path="/var/lib/kubelet/pods/215dfd5c-6d06-42ef-89c7-6d40910fbf28/volumes" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.394055 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="250d3d1c-a665-4e77-8ac0-15768f6a25f1" path="/var/lib/kubelet/pods/250d3d1c-a665-4e77-8ac0-15768f6a25f1/volumes" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.395075 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2afe8078-9724-4ab3-b5ee-79c978ce9bea" path="/var/lib/kubelet/pods/2afe8078-9724-4ab3-b5ee-79c978ce9bea/volumes" Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.498692 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.500877 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p"] Mar 09 18:57:51 crc kubenswrapper[4750]: I0309 18:57:51.733603 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" event={"ID":"ea9561e1-8a07-4d8a-9726-d1e06b93af82","Type":"ContainerStarted","Data":"2d364ff6b931cce2427dabf3cacd2f281ea1a790fda0ae62263beafa1674a2e3"} Mar 09 18:57:52 crc kubenswrapper[4750]: I0309 18:57:52.744875 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" event={"ID":"ea9561e1-8a07-4d8a-9726-d1e06b93af82","Type":"ContainerStarted","Data":"bb59f282858fd8f87208f3aa57c6e00c98695e4a9ea21186adddbb10fa5ba159"} Mar 09 18:57:52 crc kubenswrapper[4750]: I0309 18:57:52.773300 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" podStartSLOduration=2.313237953 podStartE2EDuration="2.773280904s" podCreationTimestamp="2026-03-09 18:57:50 +0000 UTC" firstStartedPulling="2026-03-09 18:57:51.498392345 +0000 UTC m=+1952.840864743" lastFinishedPulling="2026-03-09 18:57:51.958435286 +0000 UTC m=+1953.300907694" observedRunningTime="2026-03-09 18:57:52.765380229 +0000 UTC m=+1954.107852637" watchObservedRunningTime="2026-03-09 18:57:52.773280904 +0000 UTC m=+1954.115753302" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.144718 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551378-tm6ft"] Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.146972 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551378-tm6ft" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.155690 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.156047 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.156130 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.186083 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551378-tm6ft"] Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.243308 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkpfj\" (UniqueName: \"kubernetes.io/projected/004b9dfd-4abd-449e-87f1-a665e958db40-kube-api-access-dkpfj\") pod \"auto-csr-approver-29551378-tm6ft\" (UID: \"004b9dfd-4abd-449e-87f1-a665e958db40\") " pod="openshift-infra/auto-csr-approver-29551378-tm6ft" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.345492 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkpfj\" (UniqueName: \"kubernetes.io/projected/004b9dfd-4abd-449e-87f1-a665e958db40-kube-api-access-dkpfj\") pod \"auto-csr-approver-29551378-tm6ft\" (UID: \"004b9dfd-4abd-449e-87f1-a665e958db40\") " pod="openshift-infra/auto-csr-approver-29551378-tm6ft" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.372922 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkpfj\" (UniqueName: \"kubernetes.io/projected/004b9dfd-4abd-449e-87f1-a665e958db40-kube-api-access-dkpfj\") pod \"auto-csr-approver-29551378-tm6ft\" (UID: \"004b9dfd-4abd-449e-87f1-a665e958db40\") " pod="openshift-infra/auto-csr-approver-29551378-tm6ft" Mar 09 18:58:00 crc kubenswrapper[4750]: I0309 18:58:00.497684 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551378-tm6ft" Mar 09 18:58:01 crc kubenswrapper[4750]: I0309 18:58:01.008891 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551378-tm6ft"] Mar 09 18:58:01 crc kubenswrapper[4750]: I0309 18:58:01.876685 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551378-tm6ft" event={"ID":"004b9dfd-4abd-449e-87f1-a665e958db40","Type":"ContainerStarted","Data":"723c4c07d417034ef17597ccb772e2ab17177af9b9e023e7f03d0c3f017668f4"} Mar 09 18:58:02 crc kubenswrapper[4750]: I0309 18:58:02.885342 4750 generic.go:334] "Generic (PLEG): container finished" podID="004b9dfd-4abd-449e-87f1-a665e958db40" containerID="cb3b6f5397daa45f40f83323b70054bd4567857753ee307e3e6ca6a1fce62660" exitCode=0 Mar 09 18:58:02 crc kubenswrapper[4750]: I0309 18:58:02.885415 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551378-tm6ft" event={"ID":"004b9dfd-4abd-449e-87f1-a665e958db40","Type":"ContainerDied","Data":"cb3b6f5397daa45f40f83323b70054bd4567857753ee307e3e6ca6a1fce62660"} Mar 09 18:58:03 crc kubenswrapper[4750]: I0309 18:58:03.374606 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:58:03 crc kubenswrapper[4750]: E0309 18:58:03.375589 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:58:04 crc kubenswrapper[4750]: I0309 18:58:04.348541 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551378-tm6ft" Mar 09 18:58:04 crc kubenswrapper[4750]: I0309 18:58:04.535997 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkpfj\" (UniqueName: \"kubernetes.io/projected/004b9dfd-4abd-449e-87f1-a665e958db40-kube-api-access-dkpfj\") pod \"004b9dfd-4abd-449e-87f1-a665e958db40\" (UID: \"004b9dfd-4abd-449e-87f1-a665e958db40\") " Mar 09 18:58:04 crc kubenswrapper[4750]: I0309 18:58:04.545509 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/004b9dfd-4abd-449e-87f1-a665e958db40-kube-api-access-dkpfj" (OuterVolumeSpecName: "kube-api-access-dkpfj") pod "004b9dfd-4abd-449e-87f1-a665e958db40" (UID: "004b9dfd-4abd-449e-87f1-a665e958db40"). InnerVolumeSpecName "kube-api-access-dkpfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:58:04 crc kubenswrapper[4750]: I0309 18:58:04.640084 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkpfj\" (UniqueName: \"kubernetes.io/projected/004b9dfd-4abd-449e-87f1-a665e958db40-kube-api-access-dkpfj\") on node \"crc\" DevicePath \"\"" Mar 09 18:58:04 crc kubenswrapper[4750]: I0309 18:58:04.924083 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551378-tm6ft" event={"ID":"004b9dfd-4abd-449e-87f1-a665e958db40","Type":"ContainerDied","Data":"723c4c07d417034ef17597ccb772e2ab17177af9b9e023e7f03d0c3f017668f4"} Mar 09 18:58:04 crc kubenswrapper[4750]: I0309 18:58:04.924141 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="723c4c07d417034ef17597ccb772e2ab17177af9b9e023e7f03d0c3f017668f4" Mar 09 18:58:04 crc kubenswrapper[4750]: I0309 18:58:04.924223 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551378-tm6ft" Mar 09 18:58:05 crc kubenswrapper[4750]: I0309 18:58:05.438161 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551372-tbfv8"] Mar 09 18:58:05 crc kubenswrapper[4750]: I0309 18:58:05.459124 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551372-tbfv8"] Mar 09 18:58:06 crc kubenswrapper[4750]: I0309 18:58:06.046821 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-89k9z"] Mar 09 18:58:06 crc kubenswrapper[4750]: I0309 18:58:06.066288 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-89k9z"] Mar 09 18:58:07 crc kubenswrapper[4750]: I0309 18:58:07.393266 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1499c34-8025-4162-a499-662676c5e447" path="/var/lib/kubelet/pods/e1499c34-8025-4162-a499-662676c5e447/volumes" Mar 09 18:58:07 crc kubenswrapper[4750]: I0309 18:58:07.394541 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fec39559-ddae-4e19-8941-0fc3209084cd" path="/var/lib/kubelet/pods/fec39559-ddae-4e19-8941-0fc3209084cd/volumes" Mar 09 18:58:15 crc kubenswrapper[4750]: I0309 18:58:15.374216 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:58:15 crc kubenswrapper[4750]: E0309 18:58:15.375137 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:58:30 crc kubenswrapper[4750]: I0309 18:58:30.373463 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:58:30 crc kubenswrapper[4750]: E0309 18:58:30.374339 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:58:41 crc kubenswrapper[4750]: I0309 18:58:41.374124 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:58:41 crc kubenswrapper[4750]: E0309 18:58:41.375246 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:58:43 crc kubenswrapper[4750]: I0309 18:58:43.685346 4750 scope.go:117] "RemoveContainer" containerID="fa00c999ec13985a6d12a0ee552edbc27058d1f0a0de89d2e82c9c53574c40f7" Mar 09 18:58:43 crc kubenswrapper[4750]: I0309 18:58:43.744680 4750 scope.go:117] "RemoveContainer" containerID="7ccb2dde1714b67016d32aa2bf42313d9a79e120ae7ad1e8ed5ac4e14697f4ea" Mar 09 18:58:43 crc kubenswrapper[4750]: I0309 18:58:43.782787 4750 scope.go:117] "RemoveContainer" containerID="4dce597ebad76d251e228f192c4d14f53450ecec2148f3ef5d696c5786590fe4" Mar 09 18:58:43 crc kubenswrapper[4750]: I0309 18:58:43.856788 4750 scope.go:117] "RemoveContainer" containerID="e0b6fa7f3f62deb21226eab9359b6cbd960d2e58ccd84ff89b6776767266058e" Mar 09 18:58:43 crc kubenswrapper[4750]: I0309 18:58:43.900910 4750 scope.go:117] "RemoveContainer" containerID="8e3d7cd0c864c75c3c552a0a15ab5245d401dcf5c5c8c397a62a26b48376969d" Mar 09 18:58:44 crc kubenswrapper[4750]: I0309 18:58:44.062745 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-08df-account-create-update-9t8fq"] Mar 09 18:58:44 crc kubenswrapper[4750]: I0309 18:58:44.073985 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dk8lh"] Mar 09 18:58:44 crc kubenswrapper[4750]: I0309 18:58:44.084168 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dk8lh"] Mar 09 18:58:44 crc kubenswrapper[4750]: I0309 18:58:44.091973 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-08df-account-create-update-9t8fq"] Mar 09 18:58:45 crc kubenswrapper[4750]: I0309 18:58:45.039569 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b786-account-create-update-dr9fd"] Mar 09 18:58:45 crc kubenswrapper[4750]: I0309 18:58:45.053153 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b786-account-create-update-dr9fd"] Mar 09 18:58:45 crc kubenswrapper[4750]: I0309 18:58:45.382779 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081630cf-500b-49f1-992d-233ed78ff1bc" path="/var/lib/kubelet/pods/081630cf-500b-49f1-992d-233ed78ff1bc/volumes" Mar 09 18:58:45 crc kubenswrapper[4750]: I0309 18:58:45.383379 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b2776f-8175-48d4-8c86-9269df88aa64" path="/var/lib/kubelet/pods/e3b2776f-8175-48d4-8c86-9269df88aa64/volumes" Mar 09 18:58:45 crc kubenswrapper[4750]: I0309 18:58:45.384042 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9746db5-1458-4326-ba1f-5b3abbecce78" path="/var/lib/kubelet/pods/f9746db5-1458-4326-ba1f-5b3abbecce78/volumes" Mar 09 18:58:46 crc kubenswrapper[4750]: I0309 18:58:46.048263 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-qn2pj"] Mar 09 18:58:46 crc kubenswrapper[4750]: I0309 18:58:46.058536 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5f67-account-create-update-l2t7v"] Mar 09 18:58:46 crc kubenswrapper[4750]: I0309 18:58:46.067075 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-j4pfx"] Mar 09 18:58:46 crc kubenswrapper[4750]: I0309 18:58:46.077187 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-qn2pj"] Mar 09 18:58:46 crc kubenswrapper[4750]: I0309 18:58:46.084926 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-j4pfx"] Mar 09 18:58:46 crc kubenswrapper[4750]: I0309 18:58:46.092874 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5f67-account-create-update-l2t7v"] Mar 09 18:58:47 crc kubenswrapper[4750]: I0309 18:58:47.392340 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32248ec3-5f60-4ded-b77d-b9e88b7dcfe6" path="/var/lib/kubelet/pods/32248ec3-5f60-4ded-b77d-b9e88b7dcfe6/volumes" Mar 09 18:58:47 crc kubenswrapper[4750]: I0309 18:58:47.393278 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55185786-16ac-45a0-90b2-e44d0e833c23" path="/var/lib/kubelet/pods/55185786-16ac-45a0-90b2-e44d0e833c23/volumes" Mar 09 18:58:47 crc kubenswrapper[4750]: I0309 18:58:47.394121 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3837950-d435-4f57-9bde-3c5172cba1d8" path="/var/lib/kubelet/pods/e3837950-d435-4f57-9bde-3c5172cba1d8/volumes" Mar 09 18:58:52 crc kubenswrapper[4750]: I0309 18:58:52.373846 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:58:52 crc kubenswrapper[4750]: E0309 18:58:52.374639 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:59:01 crc kubenswrapper[4750]: I0309 18:59:01.108617 4750 generic.go:334] "Generic (PLEG): container finished" podID="ea9561e1-8a07-4d8a-9726-d1e06b93af82" containerID="bb59f282858fd8f87208f3aa57c6e00c98695e4a9ea21186adddbb10fa5ba159" exitCode=0 Mar 09 18:59:01 crc kubenswrapper[4750]: I0309 18:59:01.108709 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" event={"ID":"ea9561e1-8a07-4d8a-9726-d1e06b93af82","Type":"ContainerDied","Data":"bb59f282858fd8f87208f3aa57c6e00c98695e4a9ea21186adddbb10fa5ba159"} Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.663595 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.692599 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-ssh-key-openstack-edpm-ipam\") pod \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.692715 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-inventory\") pod \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.692970 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77jjv\" (UniqueName: \"kubernetes.io/projected/ea9561e1-8a07-4d8a-9726-d1e06b93af82-kube-api-access-77jjv\") pod \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\" (UID: \"ea9561e1-8a07-4d8a-9726-d1e06b93af82\") " Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.707975 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9561e1-8a07-4d8a-9726-d1e06b93af82-kube-api-access-77jjv" (OuterVolumeSpecName: "kube-api-access-77jjv") pod "ea9561e1-8a07-4d8a-9726-d1e06b93af82" (UID: "ea9561e1-8a07-4d8a-9726-d1e06b93af82"). InnerVolumeSpecName "kube-api-access-77jjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.722473 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-inventory" (OuterVolumeSpecName: "inventory") pod "ea9561e1-8a07-4d8a-9726-d1e06b93af82" (UID: "ea9561e1-8a07-4d8a-9726-d1e06b93af82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.723915 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ea9561e1-8a07-4d8a-9726-d1e06b93af82" (UID: "ea9561e1-8a07-4d8a-9726-d1e06b93af82"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.795125 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77jjv\" (UniqueName: \"kubernetes.io/projected/ea9561e1-8a07-4d8a-9726-d1e06b93af82-kube-api-access-77jjv\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.795155 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:02 crc kubenswrapper[4750]: I0309 18:59:02.795167 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea9561e1-8a07-4d8a-9726-d1e06b93af82-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.137567 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" event={"ID":"ea9561e1-8a07-4d8a-9726-d1e06b93af82","Type":"ContainerDied","Data":"2d364ff6b931cce2427dabf3cacd2f281ea1a790fda0ae62263beafa1674a2e3"} Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.137608 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d364ff6b931cce2427dabf3cacd2f281ea1a790fda0ae62263beafa1674a2e3" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.138155 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7t27p" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.273906 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5"] Mar 09 18:59:03 crc kubenswrapper[4750]: E0309 18:59:03.274431 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9561e1-8a07-4d8a-9726-d1e06b93af82" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.274461 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9561e1-8a07-4d8a-9726-d1e06b93af82" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:03 crc kubenswrapper[4750]: E0309 18:59:03.274481 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="004b9dfd-4abd-449e-87f1-a665e958db40" containerName="oc" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.274489 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="004b9dfd-4abd-449e-87f1-a665e958db40" containerName="oc" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.274760 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="004b9dfd-4abd-449e-87f1-a665e958db40" containerName="oc" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.274786 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea9561e1-8a07-4d8a-9726-d1e06b93af82" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.275662 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.280920 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.281439 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.281719 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.282041 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.291338 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5"] Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.317294 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.317365 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqjfj\" (UniqueName: \"kubernetes.io/projected/dff3602d-ae54-49a1-8d52-b7eb251a680e-kube-api-access-hqjfj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.317435 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.420004 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.420346 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.420430 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqjfj\" (UniqueName: \"kubernetes.io/projected/dff3602d-ae54-49a1-8d52-b7eb251a680e-kube-api-access-hqjfj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.425136 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.425328 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.450554 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqjfj\" (UniqueName: \"kubernetes.io/projected/dff3602d-ae54-49a1-8d52-b7eb251a680e-kube-api-access-hqjfj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:03 crc kubenswrapper[4750]: I0309 18:59:03.596305 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:04 crc kubenswrapper[4750]: I0309 18:59:04.186047 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5"] Mar 09 18:59:05 crc kubenswrapper[4750]: I0309 18:59:05.161736 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" event={"ID":"dff3602d-ae54-49a1-8d52-b7eb251a680e","Type":"ContainerStarted","Data":"c6c3f18d93370a565bffd928c924d5fd1e33c265b06bbb730a688a302a8e4520"} Mar 09 18:59:05 crc kubenswrapper[4750]: I0309 18:59:05.162540 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" event={"ID":"dff3602d-ae54-49a1-8d52-b7eb251a680e","Type":"ContainerStarted","Data":"c2150db550a8badc3c2385abeded3e02659ffa57446b8fb70791810cefa7c0ea"} Mar 09 18:59:05 crc kubenswrapper[4750]: I0309 18:59:05.183404 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" podStartSLOduration=1.770618035 podStartE2EDuration="2.183386076s" podCreationTimestamp="2026-03-09 18:59:03 +0000 UTC" firstStartedPulling="2026-03-09 18:59:04.193704816 +0000 UTC m=+2025.536177214" lastFinishedPulling="2026-03-09 18:59:04.606472857 +0000 UTC m=+2025.948945255" observedRunningTime="2026-03-09 18:59:05.175073789 +0000 UTC m=+2026.517546197" watchObservedRunningTime="2026-03-09 18:59:05.183386076 +0000 UTC m=+2026.525858474" Mar 09 18:59:07 crc kubenswrapper[4750]: I0309 18:59:07.373961 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:59:07 crc kubenswrapper[4750]: E0309 18:59:07.374797 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:59:10 crc kubenswrapper[4750]: I0309 18:59:10.219730 4750 generic.go:334] "Generic (PLEG): container finished" podID="dff3602d-ae54-49a1-8d52-b7eb251a680e" containerID="c6c3f18d93370a565bffd928c924d5fd1e33c265b06bbb730a688a302a8e4520" exitCode=0 Mar 09 18:59:10 crc kubenswrapper[4750]: I0309 18:59:10.219989 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" event={"ID":"dff3602d-ae54-49a1-8d52-b7eb251a680e","Type":"ContainerDied","Data":"c6c3f18d93370a565bffd928c924d5fd1e33c265b06bbb730a688a302a8e4520"} Mar 09 18:59:11 crc kubenswrapper[4750]: I0309 18:59:11.707733 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:11 crc kubenswrapper[4750]: I0309 18:59:11.905304 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-inventory\") pod \"dff3602d-ae54-49a1-8d52-b7eb251a680e\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " Mar 09 18:59:11 crc kubenswrapper[4750]: I0309 18:59:11.905349 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqjfj\" (UniqueName: \"kubernetes.io/projected/dff3602d-ae54-49a1-8d52-b7eb251a680e-kube-api-access-hqjfj\") pod \"dff3602d-ae54-49a1-8d52-b7eb251a680e\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " Mar 09 18:59:11 crc kubenswrapper[4750]: I0309 18:59:11.905433 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-ssh-key-openstack-edpm-ipam\") pod \"dff3602d-ae54-49a1-8d52-b7eb251a680e\" (UID: \"dff3602d-ae54-49a1-8d52-b7eb251a680e\") " Mar 09 18:59:11 crc kubenswrapper[4750]: I0309 18:59:11.912953 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dff3602d-ae54-49a1-8d52-b7eb251a680e-kube-api-access-hqjfj" (OuterVolumeSpecName: "kube-api-access-hqjfj") pod "dff3602d-ae54-49a1-8d52-b7eb251a680e" (UID: "dff3602d-ae54-49a1-8d52-b7eb251a680e"). InnerVolumeSpecName "kube-api-access-hqjfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:59:11 crc kubenswrapper[4750]: I0309 18:59:11.940381 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-inventory" (OuterVolumeSpecName: "inventory") pod "dff3602d-ae54-49a1-8d52-b7eb251a680e" (UID: "dff3602d-ae54-49a1-8d52-b7eb251a680e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:59:11 crc kubenswrapper[4750]: I0309 18:59:11.941121 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dff3602d-ae54-49a1-8d52-b7eb251a680e" (UID: "dff3602d-ae54-49a1-8d52-b7eb251a680e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.007451 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.007503 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqjfj\" (UniqueName: \"kubernetes.io/projected/dff3602d-ae54-49a1-8d52-b7eb251a680e-kube-api-access-hqjfj\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.007517 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dff3602d-ae54-49a1-8d52-b7eb251a680e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.265329 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" event={"ID":"dff3602d-ae54-49a1-8d52-b7eb251a680e","Type":"ContainerDied","Data":"c2150db550a8badc3c2385abeded3e02659ffa57446b8fb70791810cefa7c0ea"} Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.265835 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2150db550a8badc3c2385abeded3e02659ffa57446b8fb70791810cefa7c0ea" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.265469 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.408839 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9"] Mar 09 18:59:12 crc kubenswrapper[4750]: E0309 18:59:12.409599 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff3602d-ae54-49a1-8d52-b7eb251a680e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.409711 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff3602d-ae54-49a1-8d52-b7eb251a680e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.410024 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="dff3602d-ae54-49a1-8d52-b7eb251a680e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.410942 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.414306 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.416997 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.421169 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.421623 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.427595 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9"] Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.517856 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.518127 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t2z5\" (UniqueName: \"kubernetes.io/projected/316b2da2-4ec6-4c70-83ab-4c6b5e330032-kube-api-access-6t2z5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.518154 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.620983 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.621238 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t2z5\" (UniqueName: \"kubernetes.io/projected/316b2da2-4ec6-4c70-83ab-4c6b5e330032-kube-api-access-6t2z5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.621410 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.626589 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.630603 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.643500 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t2z5\" (UniqueName: \"kubernetes.io/projected/316b2da2-4ec6-4c70-83ab-4c6b5e330032-kube-api-access-6t2z5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xk6t9\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:12 crc kubenswrapper[4750]: I0309 18:59:12.729129 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:13 crc kubenswrapper[4750]: I0309 18:59:13.334864 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9"] Mar 09 18:59:13 crc kubenswrapper[4750]: W0309 18:59:13.341936 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod316b2da2_4ec6_4c70_83ab_4c6b5e330032.slice/crio-85a80be4dd320149b434519054036be0c5076c8127f66b58cb4fa2153b5b45de WatchSource:0}: Error finding container 85a80be4dd320149b434519054036be0c5076c8127f66b58cb4fa2153b5b45de: Status 404 returned error can't find the container with id 85a80be4dd320149b434519054036be0c5076c8127f66b58cb4fa2153b5b45de Mar 09 18:59:14 crc kubenswrapper[4750]: I0309 18:59:14.301135 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" event={"ID":"316b2da2-4ec6-4c70-83ab-4c6b5e330032","Type":"ContainerStarted","Data":"85a80be4dd320149b434519054036be0c5076c8127f66b58cb4fa2153b5b45de"} Mar 09 18:59:15 crc kubenswrapper[4750]: I0309 18:59:15.312883 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" event={"ID":"316b2da2-4ec6-4c70-83ab-4c6b5e330032","Type":"ContainerStarted","Data":"da337d752bd34141c777a0114f3d8f550d1bbf5aa2eb1759d0cd7169fa859b4b"} Mar 09 18:59:15 crc kubenswrapper[4750]: I0309 18:59:15.336196 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" podStartSLOduration=2.570584213 podStartE2EDuration="3.336165738s" podCreationTimestamp="2026-03-09 18:59:12 +0000 UTC" firstStartedPulling="2026-03-09 18:59:13.344834004 +0000 UTC m=+2034.687306442" lastFinishedPulling="2026-03-09 18:59:14.110415559 +0000 UTC m=+2035.452887967" observedRunningTime="2026-03-09 18:59:15.332413726 +0000 UTC m=+2036.674886134" watchObservedRunningTime="2026-03-09 18:59:15.336165738 +0000 UTC m=+2036.678638186" Mar 09 18:59:20 crc kubenswrapper[4750]: I0309 18:59:20.373171 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:59:20 crc kubenswrapper[4750]: E0309 18:59:20.374220 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 18:59:21 crc kubenswrapper[4750]: I0309 18:59:21.048758 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5jn6"] Mar 09 18:59:21 crc kubenswrapper[4750]: I0309 18:59:21.058957 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5jn6"] Mar 09 18:59:21 crc kubenswrapper[4750]: I0309 18:59:21.446373 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31e486b2-f819-4c3a-b7db-2262ad76bbb2" path="/var/lib/kubelet/pods/31e486b2-f819-4c3a-b7db-2262ad76bbb2/volumes" Mar 09 18:59:32 crc kubenswrapper[4750]: I0309 18:59:32.373597 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 18:59:33 crc kubenswrapper[4750]: I0309 18:59:33.577952 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"9744b7c86b0d0903700b40068ae49891b7ad365bb53b7b535d0120e11bb79ec7"} Mar 09 18:59:44 crc kubenswrapper[4750]: I0309 18:59:44.063321 4750 scope.go:117] "RemoveContainer" containerID="3628c72665cdec94060f5e21f9965583db46f67b888c64cc7af16cd84ad4c2a6" Mar 09 18:59:44 crc kubenswrapper[4750]: I0309 18:59:44.096565 4750 scope.go:117] "RemoveContainer" containerID="78cc53e3a4300b13510980d34453082e4103251802af1436543224ed626b17b8" Mar 09 18:59:44 crc kubenswrapper[4750]: I0309 18:59:44.166067 4750 scope.go:117] "RemoveContainer" containerID="556664ac95ccd0195f8e2bbd01932d6e568526ade20a2dc2eba689eda8675b68" Mar 09 18:59:44 crc kubenswrapper[4750]: I0309 18:59:44.215509 4750 scope.go:117] "RemoveContainer" containerID="9a8666f2d050f4143bc6b8c4f1b6a39f930ca1b8ba111b6cd701d0e899769dd6" Mar 09 18:59:44 crc kubenswrapper[4750]: I0309 18:59:44.263945 4750 scope.go:117] "RemoveContainer" containerID="0a9740590b05e48961606c6388f58f7ec58ddafc85b5662202bb308aca94c9d0" Mar 09 18:59:44 crc kubenswrapper[4750]: I0309 18:59:44.312811 4750 scope.go:117] "RemoveContainer" containerID="7764ae635f169cc2206c2082440203a3b76a001e64c8eab35587e01281ad3419" Mar 09 18:59:44 crc kubenswrapper[4750]: I0309 18:59:44.375419 4750 scope.go:117] "RemoveContainer" containerID="d6cb1251bdad02583497704b7ed6a8c6e3a45e2134af935927f1aad66508597f" Mar 09 18:59:52 crc kubenswrapper[4750]: I0309 18:59:52.788790 4750 generic.go:334] "Generic (PLEG): container finished" podID="316b2da2-4ec6-4c70-83ab-4c6b5e330032" containerID="da337d752bd34141c777a0114f3d8f550d1bbf5aa2eb1759d0cd7169fa859b4b" exitCode=0 Mar 09 18:59:52 crc kubenswrapper[4750]: I0309 18:59:52.788885 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" event={"ID":"316b2da2-4ec6-4c70-83ab-4c6b5e330032","Type":"ContainerDied","Data":"da337d752bd34141c777a0114f3d8f550d1bbf5aa2eb1759d0cd7169fa859b4b"} Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.308815 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.394225 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-ssh-key-openstack-edpm-ipam\") pod \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.394736 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-inventory\") pod \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.395008 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t2z5\" (UniqueName: \"kubernetes.io/projected/316b2da2-4ec6-4c70-83ab-4c6b5e330032-kube-api-access-6t2z5\") pod \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\" (UID: \"316b2da2-4ec6-4c70-83ab-4c6b5e330032\") " Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.399704 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316b2da2-4ec6-4c70-83ab-4c6b5e330032-kube-api-access-6t2z5" (OuterVolumeSpecName: "kube-api-access-6t2z5") pod "316b2da2-4ec6-4c70-83ab-4c6b5e330032" (UID: "316b2da2-4ec6-4c70-83ab-4c6b5e330032"). InnerVolumeSpecName "kube-api-access-6t2z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.425654 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-inventory" (OuterVolumeSpecName: "inventory") pod "316b2da2-4ec6-4c70-83ab-4c6b5e330032" (UID: "316b2da2-4ec6-4c70-83ab-4c6b5e330032"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.473245 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "316b2da2-4ec6-4c70-83ab-4c6b5e330032" (UID: "316b2da2-4ec6-4c70-83ab-4c6b5e330032"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.499711 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.499744 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/316b2da2-4ec6-4c70-83ab-4c6b5e330032-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.499754 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t2z5\" (UniqueName: \"kubernetes.io/projected/316b2da2-4ec6-4c70-83ab-4c6b5e330032-kube-api-access-6t2z5\") on node \"crc\" DevicePath \"\"" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.813868 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" event={"ID":"316b2da2-4ec6-4c70-83ab-4c6b5e330032","Type":"ContainerDied","Data":"85a80be4dd320149b434519054036be0c5076c8127f66b58cb4fa2153b5b45de"} Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.813911 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85a80be4dd320149b434519054036be0c5076c8127f66b58cb4fa2153b5b45de" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.813974 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xk6t9" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.926807 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5"] Mar 09 18:59:54 crc kubenswrapper[4750]: E0309 18:59:54.927307 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316b2da2-4ec6-4c70-83ab-4c6b5e330032" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.927334 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="316b2da2-4ec6-4c70-83ab-4c6b5e330032" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.927573 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="316b2da2-4ec6-4c70-83ab-4c6b5e330032" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.928501 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.931746 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.932449 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.933383 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.937536 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 18:59:54 crc kubenswrapper[4750]: I0309 18:59:54.939126 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5"] Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.009373 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.009438 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.009525 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgjdx\" (UniqueName: \"kubernetes.io/projected/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-kube-api-access-xgjdx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.111892 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.111995 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.112786 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgjdx\" (UniqueName: \"kubernetes.io/projected/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-kube-api-access-xgjdx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.115367 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.116245 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.137537 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgjdx\" (UniqueName: \"kubernetes.io/projected/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-kube-api-access-xgjdx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.246140 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 18:59:55 crc kubenswrapper[4750]: I0309 18:59:55.844969 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5"] Mar 09 18:59:56 crc kubenswrapper[4750]: I0309 18:59:56.836688 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" event={"ID":"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2","Type":"ContainerStarted","Data":"c84c8800e8b4f86c464ec5569ee5d7bbb148257f61b56928312bca65154bf0cb"} Mar 09 18:59:56 crc kubenswrapper[4750]: I0309 18:59:56.837095 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" event={"ID":"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2","Type":"ContainerStarted","Data":"33b6aa7db692139ad74ab890baf055e658db9edba7d694aa9d230cdb1ca7f8ba"} Mar 09 18:59:56 crc kubenswrapper[4750]: I0309 18:59:56.863687 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" podStartSLOduration=2.410583739 podStartE2EDuration="2.863625144s" podCreationTimestamp="2026-03-09 18:59:54 +0000 UTC" firstStartedPulling="2026-03-09 18:59:55.85989803 +0000 UTC m=+2077.202370428" lastFinishedPulling="2026-03-09 18:59:56.312939395 +0000 UTC m=+2077.655411833" observedRunningTime="2026-03-09 18:59:56.857128507 +0000 UTC m=+2078.199600905" watchObservedRunningTime="2026-03-09 18:59:56.863625144 +0000 UTC m=+2078.206097572" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.158768 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551380-5m6ww"] Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.161666 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551380-5m6ww" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.164695 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.164807 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.165246 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.179233 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm"] Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.181067 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.184931 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.186725 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.209848 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551380-5m6ww"] Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.221810 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm"] Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.242541 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-secret-volume\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.242862 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckspk\" (UniqueName: \"kubernetes.io/projected/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-kube-api-access-ckspk\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.242953 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-config-volume\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.243137 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfc9f\" (UniqueName: \"kubernetes.io/projected/2a135e52-9627-4241-b557-a5146429cdec-kube-api-access-zfc9f\") pod \"auto-csr-approver-29551380-5m6ww\" (UID: \"2a135e52-9627-4241-b557-a5146429cdec\") " pod="openshift-infra/auto-csr-approver-29551380-5m6ww" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.344617 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-secret-volume\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.344830 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckspk\" (UniqueName: \"kubernetes.io/projected/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-kube-api-access-ckspk\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.344868 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-config-volume\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.344941 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfc9f\" (UniqueName: \"kubernetes.io/projected/2a135e52-9627-4241-b557-a5146429cdec-kube-api-access-zfc9f\") pod \"auto-csr-approver-29551380-5m6ww\" (UID: \"2a135e52-9627-4241-b557-a5146429cdec\") " pod="openshift-infra/auto-csr-approver-29551380-5m6ww" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.346371 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-config-volume\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.354491 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-secret-volume\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.364032 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfc9f\" (UniqueName: \"kubernetes.io/projected/2a135e52-9627-4241-b557-a5146429cdec-kube-api-access-zfc9f\") pod \"auto-csr-approver-29551380-5m6ww\" (UID: \"2a135e52-9627-4241-b557-a5146429cdec\") " pod="openshift-infra/auto-csr-approver-29551380-5m6ww" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.367891 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckspk\" (UniqueName: \"kubernetes.io/projected/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-kube-api-access-ckspk\") pod \"collect-profiles-29551380-h4ckm\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.492234 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551380-5m6ww" Mar 09 19:00:00 crc kubenswrapper[4750]: I0309 19:00:00.522692 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:01 crc kubenswrapper[4750]: W0309 19:00:01.063329 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a135e52_9627_4241_b557_a5146429cdec.slice/crio-1ed959bae064cfd02c2c7ee88c7abf21668dbed3ffbaa7d834e86555a6c8f949 WatchSource:0}: Error finding container 1ed959bae064cfd02c2c7ee88c7abf21668dbed3ffbaa7d834e86555a6c8f949: Status 404 returned error can't find the container with id 1ed959bae064cfd02c2c7ee88c7abf21668dbed3ffbaa7d834e86555a6c8f949 Mar 09 19:00:01 crc kubenswrapper[4750]: I0309 19:00:01.069764 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551380-5m6ww"] Mar 09 19:00:01 crc kubenswrapper[4750]: I0309 19:00:01.159575 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm"] Mar 09 19:00:01 crc kubenswrapper[4750]: W0309 19:00:01.167235 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb726a609_8d2b_406e_8f7a_1e37d2c8a7a5.slice/crio-270b4a7ca8b9472569e7fa9628f71c0a192d3ab0fa532d93cba2180de2ec83d5 WatchSource:0}: Error finding container 270b4a7ca8b9472569e7fa9628f71c0a192d3ab0fa532d93cba2180de2ec83d5: Status 404 returned error can't find the container with id 270b4a7ca8b9472569e7fa9628f71c0a192d3ab0fa532d93cba2180de2ec83d5 Mar 09 19:00:01 crc kubenswrapper[4750]: I0309 19:00:01.893581 4750 generic.go:334] "Generic (PLEG): container finished" podID="b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" containerID="3e0918883adb7aa11f8553ebdd0cf2413b90ef754e1bffcbc4c47b2b9c38dc75" exitCode=0 Mar 09 19:00:01 crc kubenswrapper[4750]: I0309 19:00:01.893689 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" event={"ID":"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5","Type":"ContainerDied","Data":"3e0918883adb7aa11f8553ebdd0cf2413b90ef754e1bffcbc4c47b2b9c38dc75"} Mar 09 19:00:01 crc kubenswrapper[4750]: I0309 19:00:01.893725 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" event={"ID":"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5","Type":"ContainerStarted","Data":"270b4a7ca8b9472569e7fa9628f71c0a192d3ab0fa532d93cba2180de2ec83d5"} Mar 09 19:00:01 crc kubenswrapper[4750]: I0309 19:00:01.895335 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551380-5m6ww" event={"ID":"2a135e52-9627-4241-b557-a5146429cdec","Type":"ContainerStarted","Data":"1ed959bae064cfd02c2c7ee88c7abf21668dbed3ffbaa7d834e86555a6c8f949"} Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.285188 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.318810 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckspk\" (UniqueName: \"kubernetes.io/projected/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-kube-api-access-ckspk\") pod \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.319041 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-config-volume\") pod \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.319154 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-secret-volume\") pod \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\" (UID: \"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5\") " Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.320086 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-config-volume" (OuterVolumeSpecName: "config-volume") pod "b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" (UID: "b726a609-8d2b-406e-8f7a-1e37d2c8a7a5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.355343 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" (UID: "b726a609-8d2b-406e-8f7a-1e37d2c8a7a5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.357930 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-kube-api-access-ckspk" (OuterVolumeSpecName: "kube-api-access-ckspk") pod "b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" (UID: "b726a609-8d2b-406e-8f7a-1e37d2c8a7a5"). InnerVolumeSpecName "kube-api-access-ckspk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.421596 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.421658 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.421672 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckspk\" (UniqueName: \"kubernetes.io/projected/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5-kube-api-access-ckspk\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.918389 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" event={"ID":"b726a609-8d2b-406e-8f7a-1e37d2c8a7a5","Type":"ContainerDied","Data":"270b4a7ca8b9472569e7fa9628f71c0a192d3ab0fa532d93cba2180de2ec83d5"} Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.918439 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="270b4a7ca8b9472569e7fa9628f71c0a192d3ab0fa532d93cba2180de2ec83d5" Mar 09 19:00:03 crc kubenswrapper[4750]: I0309 19:00:03.918476 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm" Mar 09 19:00:04 crc kubenswrapper[4750]: I0309 19:00:04.370739 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr"] Mar 09 19:00:04 crc kubenswrapper[4750]: I0309 19:00:04.380148 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551335-l5ldr"] Mar 09 19:00:04 crc kubenswrapper[4750]: I0309 19:00:04.942668 4750 generic.go:334] "Generic (PLEG): container finished" podID="2a135e52-9627-4241-b557-a5146429cdec" containerID="83f382254d36cc8e265c3d941d6a1e857ee16c6492a8ccbf68b8f8e3d3f282f6" exitCode=0 Mar 09 19:00:04 crc kubenswrapper[4750]: I0309 19:00:04.942755 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551380-5m6ww" event={"ID":"2a135e52-9627-4241-b557-a5146429cdec","Type":"ContainerDied","Data":"83f382254d36cc8e265c3d941d6a1e857ee16c6492a8ccbf68b8f8e3d3f282f6"} Mar 09 19:00:05 crc kubenswrapper[4750]: I0309 19:00:05.395474 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ecde8f-d9fa-45e5-a508-981772436b4f" path="/var/lib/kubelet/pods/d0ecde8f-d9fa-45e5-a508-981772436b4f/volumes" Mar 09 19:00:06 crc kubenswrapper[4750]: I0309 19:00:06.362314 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551380-5m6ww" Mar 09 19:00:06 crc kubenswrapper[4750]: I0309 19:00:06.400504 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfc9f\" (UniqueName: \"kubernetes.io/projected/2a135e52-9627-4241-b557-a5146429cdec-kube-api-access-zfc9f\") pod \"2a135e52-9627-4241-b557-a5146429cdec\" (UID: \"2a135e52-9627-4241-b557-a5146429cdec\") " Mar 09 19:00:06 crc kubenswrapper[4750]: I0309 19:00:06.422226 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a135e52-9627-4241-b557-a5146429cdec-kube-api-access-zfc9f" (OuterVolumeSpecName: "kube-api-access-zfc9f") pod "2a135e52-9627-4241-b557-a5146429cdec" (UID: "2a135e52-9627-4241-b557-a5146429cdec"). InnerVolumeSpecName "kube-api-access-zfc9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:00:06 crc kubenswrapper[4750]: I0309 19:00:06.503274 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfc9f\" (UniqueName: \"kubernetes.io/projected/2a135e52-9627-4241-b557-a5146429cdec-kube-api-access-zfc9f\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:06 crc kubenswrapper[4750]: I0309 19:00:06.995907 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551380-5m6ww" event={"ID":"2a135e52-9627-4241-b557-a5146429cdec","Type":"ContainerDied","Data":"1ed959bae064cfd02c2c7ee88c7abf21668dbed3ffbaa7d834e86555a6c8f949"} Mar 09 19:00:06 crc kubenswrapper[4750]: I0309 19:00:06.995975 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ed959bae064cfd02c2c7ee88c7abf21668dbed3ffbaa7d834e86555a6c8f949" Mar 09 19:00:06 crc kubenswrapper[4750]: I0309 19:00:06.995986 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551380-5m6ww" Mar 09 19:00:07 crc kubenswrapper[4750]: I0309 19:00:07.433574 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551374-w4k5b"] Mar 09 19:00:07 crc kubenswrapper[4750]: I0309 19:00:07.441460 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551374-w4k5b"] Mar 09 19:00:09 crc kubenswrapper[4750]: I0309 19:00:09.389268 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6200983-07bd-48a6-af4c-2bbfa63afdb2" path="/var/lib/kubelet/pods/f6200983-07bd-48a6-af4c-2bbfa63afdb2/volumes" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.840814 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ntlz2"] Mar 09 19:00:10 crc kubenswrapper[4750]: E0309 19:00:10.841913 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" containerName="collect-profiles" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.841938 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" containerName="collect-profiles" Mar 09 19:00:10 crc kubenswrapper[4750]: E0309 19:00:10.841960 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a135e52-9627-4241-b557-a5146429cdec" containerName="oc" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.841973 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a135e52-9627-4241-b557-a5146429cdec" containerName="oc" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.842297 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a135e52-9627-4241-b557-a5146429cdec" containerName="oc" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.842351 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" containerName="collect-profiles" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.845752 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.867232 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ntlz2"] Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.901565 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-utilities\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.901648 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-catalog-content\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:10 crc kubenswrapper[4750]: I0309 19:00:10.901751 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7fw5\" (UniqueName: \"kubernetes.io/projected/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-kube-api-access-k7fw5\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.003572 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7fw5\" (UniqueName: \"kubernetes.io/projected/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-kube-api-access-k7fw5\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.003737 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-utilities\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.003759 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-catalog-content\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.004260 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-catalog-content\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.004361 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-utilities\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.028617 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7fw5\" (UniqueName: \"kubernetes.io/projected/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-kube-api-access-k7fw5\") pod \"redhat-operators-ntlz2\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.191951 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:11 crc kubenswrapper[4750]: I0309 19:00:11.740219 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ntlz2"] Mar 09 19:00:12 crc kubenswrapper[4750]: I0309 19:00:12.046263 4750 generic.go:334] "Generic (PLEG): container finished" podID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerID="ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e" exitCode=0 Mar 09 19:00:12 crc kubenswrapper[4750]: I0309 19:00:12.046469 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntlz2" event={"ID":"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1","Type":"ContainerDied","Data":"ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e"} Mar 09 19:00:12 crc kubenswrapper[4750]: I0309 19:00:12.046772 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntlz2" event={"ID":"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1","Type":"ContainerStarted","Data":"5c60033e80fb9ea26481c5259a31fbf9697b95b4472be92095077cb18b4d9c06"} Mar 09 19:00:14 crc kubenswrapper[4750]: I0309 19:00:14.088455 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntlz2" event={"ID":"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1","Type":"ContainerStarted","Data":"5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890"} Mar 09 19:00:20 crc kubenswrapper[4750]: I0309 19:00:20.045428 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vznp2"] Mar 09 19:00:20 crc kubenswrapper[4750]: I0309 19:00:20.063161 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vznp2"] Mar 09 19:00:20 crc kubenswrapper[4750]: I0309 19:00:20.156887 4750 generic.go:334] "Generic (PLEG): container finished" podID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerID="5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890" exitCode=0 Mar 09 19:00:20 crc kubenswrapper[4750]: I0309 19:00:20.156964 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntlz2" event={"ID":"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1","Type":"ContainerDied","Data":"5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890"} Mar 09 19:00:21 crc kubenswrapper[4750]: I0309 19:00:21.386178 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df584b8-cb6c-4044-a738-8500b2a3e78b" path="/var/lib/kubelet/pods/7df584b8-cb6c-4044-a738-8500b2a3e78b/volumes" Mar 09 19:00:22 crc kubenswrapper[4750]: I0309 19:00:22.044499 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l99gb"] Mar 09 19:00:22 crc kubenswrapper[4750]: I0309 19:00:22.059821 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l99gb"] Mar 09 19:00:22 crc kubenswrapper[4750]: I0309 19:00:22.184181 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntlz2" event={"ID":"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1","Type":"ContainerStarted","Data":"3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf"} Mar 09 19:00:22 crc kubenswrapper[4750]: I0309 19:00:22.227469 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ntlz2" podStartSLOduration=3.3622985930000002 podStartE2EDuration="12.227429768s" podCreationTimestamp="2026-03-09 19:00:10 +0000 UTC" firstStartedPulling="2026-03-09 19:00:12.048707693 +0000 UTC m=+2093.391180111" lastFinishedPulling="2026-03-09 19:00:20.913838858 +0000 UTC m=+2102.256311286" observedRunningTime="2026-03-09 19:00:22.213047766 +0000 UTC m=+2103.555520184" watchObservedRunningTime="2026-03-09 19:00:22.227429768 +0000 UTC m=+2103.569902206" Mar 09 19:00:23 crc kubenswrapper[4750]: I0309 19:00:23.392762 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0" path="/var/lib/kubelet/pods/6fa7e3dd-d07a-48ed-8502-ef44fc7dcaa0/volumes" Mar 09 19:00:31 crc kubenswrapper[4750]: I0309 19:00:31.192342 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:31 crc kubenswrapper[4750]: I0309 19:00:31.193254 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:32 crc kubenswrapper[4750]: I0309 19:00:32.253163 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ntlz2" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="registry-server" probeResult="failure" output=< Mar 09 19:00:32 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:00:32 crc kubenswrapper[4750]: > Mar 09 19:00:41 crc kubenswrapper[4750]: I0309 19:00:41.249444 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:41 crc kubenswrapper[4750]: I0309 19:00:41.304109 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:42 crc kubenswrapper[4750]: I0309 19:00:42.035011 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ntlz2"] Mar 09 19:00:42 crc kubenswrapper[4750]: I0309 19:00:42.403968 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ntlz2" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="registry-server" containerID="cri-o://3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf" gracePeriod=2 Mar 09 19:00:42 crc kubenswrapper[4750]: I0309 19:00:42.910553 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.073162 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7fw5\" (UniqueName: \"kubernetes.io/projected/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-kube-api-access-k7fw5\") pod \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.073321 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-catalog-content\") pod \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.073372 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-utilities\") pod \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\" (UID: \"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1\") " Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.074391 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-utilities" (OuterVolumeSpecName: "utilities") pod "e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" (UID: "e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.078787 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-kube-api-access-k7fw5" (OuterVolumeSpecName: "kube-api-access-k7fw5") pod "e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" (UID: "e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1"). InnerVolumeSpecName "kube-api-access-k7fw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.176204 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7fw5\" (UniqueName: \"kubernetes.io/projected/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-kube-api-access-k7fw5\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.176471 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.215952 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" (UID: "e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.278835 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.437668 4750 generic.go:334] "Generic (PLEG): container finished" podID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerID="3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf" exitCode=0 Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.438657 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntlz2" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.438693 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntlz2" event={"ID":"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1","Type":"ContainerDied","Data":"3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf"} Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.439934 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntlz2" event={"ID":"e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1","Type":"ContainerDied","Data":"5c60033e80fb9ea26481c5259a31fbf9697b95b4472be92095077cb18b4d9c06"} Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.440018 4750 scope.go:117] "RemoveContainer" containerID="3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.476813 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ntlz2"] Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.491110 4750 scope.go:117] "RemoveContainer" containerID="5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.509780 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ntlz2"] Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.523987 4750 scope.go:117] "RemoveContainer" containerID="ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.568578 4750 scope.go:117] "RemoveContainer" containerID="3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf" Mar 09 19:00:43 crc kubenswrapper[4750]: E0309 19:00:43.569541 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf\": container with ID starting with 3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf not found: ID does not exist" containerID="3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.569670 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf"} err="failed to get container status \"3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf\": rpc error: code = NotFound desc = could not find container \"3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf\": container with ID starting with 3902265abed30fd6b2c3e2f4db750f3cae1e2f06824742cb6c206688bdc811bf not found: ID does not exist" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.569771 4750 scope.go:117] "RemoveContainer" containerID="5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890" Mar 09 19:00:43 crc kubenswrapper[4750]: E0309 19:00:43.570189 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890\": container with ID starting with 5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890 not found: ID does not exist" containerID="5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.570275 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890"} err="failed to get container status \"5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890\": rpc error: code = NotFound desc = could not find container \"5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890\": container with ID starting with 5ae6710a60b5fbb919ee8069f4ce2a347f5780589f72386f77893af0f2e73890 not found: ID does not exist" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.570345 4750 scope.go:117] "RemoveContainer" containerID="ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e" Mar 09 19:00:43 crc kubenswrapper[4750]: E0309 19:00:43.571061 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e\": container with ID starting with ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e not found: ID does not exist" containerID="ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e" Mar 09 19:00:43 crc kubenswrapper[4750]: I0309 19:00:43.571148 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e"} err="failed to get container status \"ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e\": rpc error: code = NotFound desc = could not find container \"ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e\": container with ID starting with ff1fdc295b68ebe40dba6ad36e2b03efb3a1b328fb1994f7efd44b61cc3b896e not found: ID does not exist" Mar 09 19:00:44 crc kubenswrapper[4750]: I0309 19:00:44.558272 4750 scope.go:117] "RemoveContainer" containerID="f4d56c47e3e3f39a3e1ff5e639760480fdf70839f876a79fe77b5fd26dc63c3e" Mar 09 19:00:44 crc kubenswrapper[4750]: I0309 19:00:44.630924 4750 scope.go:117] "RemoveContainer" containerID="1aa681e696281cc67b93af1e699c7c67d083d7656749096e07580423b92d0786" Mar 09 19:00:44 crc kubenswrapper[4750]: I0309 19:00:44.657114 4750 scope.go:117] "RemoveContainer" containerID="6673887b27d13a728a658ddbf971f33ab76cf69b7c031681d03daffb6b374c34" Mar 09 19:00:44 crc kubenswrapper[4750]: I0309 19:00:44.719473 4750 scope.go:117] "RemoveContainer" containerID="99e4ea5b5a58c028cf87f4fd08f6ab66a9f2db77da502f4d2b163a9c1538efad" Mar 09 19:00:45 crc kubenswrapper[4750]: I0309 19:00:45.388736 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" path="/var/lib/kubelet/pods/e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1/volumes" Mar 09 19:00:46 crc kubenswrapper[4750]: I0309 19:00:46.484659 4750 generic.go:334] "Generic (PLEG): container finished" podID="ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2" containerID="c84c8800e8b4f86c464ec5569ee5d7bbb148257f61b56928312bca65154bf0cb" exitCode=0 Mar 09 19:00:46 crc kubenswrapper[4750]: I0309 19:00:46.484720 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" event={"ID":"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2","Type":"ContainerDied","Data":"c84c8800e8b4f86c464ec5569ee5d7bbb148257f61b56928312bca65154bf0cb"} Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.022887 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.208619 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-ssh-key-openstack-edpm-ipam\") pod \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.208781 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgjdx\" (UniqueName: \"kubernetes.io/projected/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-kube-api-access-xgjdx\") pod \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.208826 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-inventory\") pod \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\" (UID: \"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2\") " Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.214486 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-kube-api-access-xgjdx" (OuterVolumeSpecName: "kube-api-access-xgjdx") pod "ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2" (UID: "ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2"). InnerVolumeSpecName "kube-api-access-xgjdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.247339 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-inventory" (OuterVolumeSpecName: "inventory") pod "ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2" (UID: "ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.276855 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2" (UID: "ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.311689 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.311736 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgjdx\" (UniqueName: \"kubernetes.io/projected/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-kube-api-access-xgjdx\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.311754 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.510523 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" event={"ID":"ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2","Type":"ContainerDied","Data":"33b6aa7db692139ad74ab890baf055e658db9edba7d694aa9d230cdb1ca7f8ba"} Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.510572 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33b6aa7db692139ad74ab890baf055e658db9edba7d694aa9d230cdb1ca7f8ba" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.510742 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.627453 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c5nfl"] Mar 09 19:00:48 crc kubenswrapper[4750]: E0309 19:00:48.627953 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="extract-utilities" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.627974 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="extract-utilities" Mar 09 19:00:48 crc kubenswrapper[4750]: E0309 19:00:48.628001 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.628010 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:00:48 crc kubenswrapper[4750]: E0309 19:00:48.628050 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="extract-content" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.628057 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="extract-content" Mar 09 19:00:48 crc kubenswrapper[4750]: E0309 19:00:48.628070 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="registry-server" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.628077 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="registry-server" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.628299 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.628314 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e12e4c9a-7a5e-4ec8-8090-5b56231fb6e1" containerName="registry-server" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.629229 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.631899 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.639894 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.639894 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.640074 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.653046 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c5nfl"] Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.820784 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.821196 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7dd8\" (UniqueName: \"kubernetes.io/projected/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-kube-api-access-h7dd8\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.821254 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.923360 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.923429 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7dd8\" (UniqueName: \"kubernetes.io/projected/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-kube-api-access-h7dd8\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.923494 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.930308 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.930395 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:48 crc kubenswrapper[4750]: I0309 19:00:48.961727 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7dd8\" (UniqueName: \"kubernetes.io/projected/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-kube-api-access-h7dd8\") pod \"ssh-known-hosts-edpm-deployment-c5nfl\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:49 crc kubenswrapper[4750]: I0309 19:00:49.253986 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:49 crc kubenswrapper[4750]: I0309 19:00:49.887430 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c5nfl"] Mar 09 19:00:49 crc kubenswrapper[4750]: W0309 19:00:49.890278 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5679eb0f_6cb5_4d30_8428_19bcf4dc13ed.slice/crio-e6e0f206ab93519f612b3ee9993d00beb17081f52ae4b4faae5cb554cc6d839c WatchSource:0}: Error finding container e6e0f206ab93519f612b3ee9993d00beb17081f52ae4b4faae5cb554cc6d839c: Status 404 returned error can't find the container with id e6e0f206ab93519f612b3ee9993d00beb17081f52ae4b4faae5cb554cc6d839c Mar 09 19:00:50 crc kubenswrapper[4750]: I0309 19:00:50.534055 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" event={"ID":"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed","Type":"ContainerStarted","Data":"e6e0f206ab93519f612b3ee9993d00beb17081f52ae4b4faae5cb554cc6d839c"} Mar 09 19:00:51 crc kubenswrapper[4750]: I0309 19:00:51.554207 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" event={"ID":"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed","Type":"ContainerStarted","Data":"4d40ad26f6b10013d56185d05c257348b99e95b4f5c88f26c91ce8e7142e4e7b"} Mar 09 19:00:57 crc kubenswrapper[4750]: I0309 19:00:57.627360 4750 generic.go:334] "Generic (PLEG): container finished" podID="5679eb0f-6cb5-4d30-8428-19bcf4dc13ed" containerID="4d40ad26f6b10013d56185d05c257348b99e95b4f5c88f26c91ce8e7142e4e7b" exitCode=0 Mar 09 19:00:57 crc kubenswrapper[4750]: I0309 19:00:57.627471 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" event={"ID":"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed","Type":"ContainerDied","Data":"4d40ad26f6b10013d56185d05c257348b99e95b4f5c88f26c91ce8e7142e4e7b"} Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.121560 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.262752 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7dd8\" (UniqueName: \"kubernetes.io/projected/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-kube-api-access-h7dd8\") pod \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.262926 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-inventory-0\") pod \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.263021 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-ssh-key-openstack-edpm-ipam\") pod \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\" (UID: \"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed\") " Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.268288 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-kube-api-access-h7dd8" (OuterVolumeSpecName: "kube-api-access-h7dd8") pod "5679eb0f-6cb5-4d30-8428-19bcf4dc13ed" (UID: "5679eb0f-6cb5-4d30-8428-19bcf4dc13ed"). InnerVolumeSpecName "kube-api-access-h7dd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.292798 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5679eb0f-6cb5-4d30-8428-19bcf4dc13ed" (UID: "5679eb0f-6cb5-4d30-8428-19bcf4dc13ed"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.307903 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5679eb0f-6cb5-4d30-8428-19bcf4dc13ed" (UID: "5679eb0f-6cb5-4d30-8428-19bcf4dc13ed"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.366188 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.366245 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7dd8\" (UniqueName: \"kubernetes.io/projected/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-kube-api-access-h7dd8\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.366268 4750 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5679eb0f-6cb5-4d30-8428-19bcf4dc13ed-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.652139 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" event={"ID":"5679eb0f-6cb5-4d30-8428-19bcf4dc13ed","Type":"ContainerDied","Data":"e6e0f206ab93519f612b3ee9993d00beb17081f52ae4b4faae5cb554cc6d839c"} Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.652418 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6e0f206ab93519f612b3ee9993d00beb17081f52ae4b4faae5cb554cc6d839c" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.652239 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c5nfl" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.766151 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s"] Mar 09 19:00:59 crc kubenswrapper[4750]: E0309 19:00:59.766785 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5679eb0f-6cb5-4d30-8428-19bcf4dc13ed" containerName="ssh-known-hosts-edpm-deployment" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.766816 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5679eb0f-6cb5-4d30-8428-19bcf4dc13ed" containerName="ssh-known-hosts-edpm-deployment" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.767236 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5679eb0f-6cb5-4d30-8428-19bcf4dc13ed" containerName="ssh-known-hosts-edpm-deployment" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.768740 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.771824 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.771976 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.771999 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.773440 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.783948 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s"] Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.878815 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94rhg\" (UniqueName: \"kubernetes.io/projected/50969409-de1c-43a2-bf66-60da87e8bcb4-kube-api-access-94rhg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.878888 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.879065 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.981645 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.981943 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.982032 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94rhg\" (UniqueName: \"kubernetes.io/projected/50969409-de1c-43a2-bf66-60da87e8bcb4-kube-api-access-94rhg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.987428 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:00:59 crc kubenswrapper[4750]: I0309 19:00:59.988441 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.002989 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94rhg\" (UniqueName: \"kubernetes.io/projected/50969409-de1c-43a2-bf66-60da87e8bcb4-kube-api-access-94rhg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-96w2s\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.109804 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.154095 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29551381-lcv2d"] Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.155945 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.188524 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29551381-lcv2d"] Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.288661 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-config-data\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.288720 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-combined-ca-bundle\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.288775 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5blxd\" (UniqueName: \"kubernetes.io/projected/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-kube-api-access-5blxd\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.288865 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-fernet-keys\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.390838 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-config-data\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.390900 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-combined-ca-bundle\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.390937 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5blxd\" (UniqueName: \"kubernetes.io/projected/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-kube-api-access-5blxd\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.390994 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-fernet-keys\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.401424 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-fernet-keys\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.405146 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-combined-ca-bundle\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.408954 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-config-data\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.419941 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5blxd\" (UniqueName: \"kubernetes.io/projected/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-kube-api-access-5blxd\") pod \"keystone-cron-29551381-lcv2d\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.560514 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.769849 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s"] Mar 09 19:01:00 crc kubenswrapper[4750]: I0309 19:01:00.845376 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29551381-lcv2d"] Mar 09 19:01:00 crc kubenswrapper[4750]: W0309 19:01:00.846479 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d707807_4c3a_4eeb_b9c5_8c785bdf3cd1.slice/crio-e472f2bfd271f214009182f3ef8734a21fb766f44d363a34993784c2b02f11ba WatchSource:0}: Error finding container e472f2bfd271f214009182f3ef8734a21fb766f44d363a34993784c2b02f11ba: Status 404 returned error can't find the container with id e472f2bfd271f214009182f3ef8734a21fb766f44d363a34993784c2b02f11ba Mar 09 19:01:01 crc kubenswrapper[4750]: I0309 19:01:01.678789 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" event={"ID":"50969409-de1c-43a2-bf66-60da87e8bcb4","Type":"ContainerStarted","Data":"d53b113d4b962200cabf36ca9ae7656f661bdddc135e576dddb995ba4c50df8e"} Mar 09 19:01:01 crc kubenswrapper[4750]: I0309 19:01:01.689277 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551381-lcv2d" event={"ID":"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1","Type":"ContainerStarted","Data":"1839261b484f0d9a5bcda531dabbacd1a082b815afbddc75ea7d81d72716e66e"} Mar 09 19:01:01 crc kubenswrapper[4750]: I0309 19:01:01.689341 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551381-lcv2d" event={"ID":"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1","Type":"ContainerStarted","Data":"e472f2bfd271f214009182f3ef8734a21fb766f44d363a34993784c2b02f11ba"} Mar 09 19:01:01 crc kubenswrapper[4750]: I0309 19:01:01.714168 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29551381-lcv2d" podStartSLOduration=1.714149122 podStartE2EDuration="1.714149122s" podCreationTimestamp="2026-03-09 19:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 19:01:01.710225675 +0000 UTC m=+2143.052698093" watchObservedRunningTime="2026-03-09 19:01:01.714149122 +0000 UTC m=+2143.056621520" Mar 09 19:01:02 crc kubenswrapper[4750]: I0309 19:01:02.050144 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-4xzzc"] Mar 09 19:01:02 crc kubenswrapper[4750]: I0309 19:01:02.058540 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-4xzzc"] Mar 09 19:01:02 crc kubenswrapper[4750]: I0309 19:01:02.706953 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" event={"ID":"50969409-de1c-43a2-bf66-60da87e8bcb4","Type":"ContainerStarted","Data":"f8d6c4ba32da9af38c8ffca9d99d2d2b9403fc44cc0bce543821275d776796db"} Mar 09 19:01:02 crc kubenswrapper[4750]: I0309 19:01:02.748774 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" podStartSLOduration=2.909394561 podStartE2EDuration="3.748744878s" podCreationTimestamp="2026-03-09 19:00:59 +0000 UTC" firstStartedPulling="2026-03-09 19:01:00.750457402 +0000 UTC m=+2142.092929830" lastFinishedPulling="2026-03-09 19:01:01.589807749 +0000 UTC m=+2142.932280147" observedRunningTime="2026-03-09 19:01:02.731350154 +0000 UTC m=+2144.073822592" watchObservedRunningTime="2026-03-09 19:01:02.748744878 +0000 UTC m=+2144.091217306" Mar 09 19:01:03 crc kubenswrapper[4750]: I0309 19:01:03.390413 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c34fb79-24fb-4f5a-b7f9-09e4ae449607" path="/var/lib/kubelet/pods/7c34fb79-24fb-4f5a-b7f9-09e4ae449607/volumes" Mar 09 19:01:03 crc kubenswrapper[4750]: I0309 19:01:03.746781 4750 generic.go:334] "Generic (PLEG): container finished" podID="1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" containerID="1839261b484f0d9a5bcda531dabbacd1a082b815afbddc75ea7d81d72716e66e" exitCode=0 Mar 09 19:01:03 crc kubenswrapper[4750]: I0309 19:01:03.746924 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551381-lcv2d" event={"ID":"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1","Type":"ContainerDied","Data":"1839261b484f0d9a5bcda531dabbacd1a082b815afbddc75ea7d81d72716e66e"} Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.154456 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.218179 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5blxd\" (UniqueName: \"kubernetes.io/projected/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-kube-api-access-5blxd\") pod \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.218311 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-config-data\") pod \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.218548 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-fernet-keys\") pod \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.218784 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-combined-ca-bundle\") pod \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\" (UID: \"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1\") " Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.228010 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" (UID: "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.233125 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-kube-api-access-5blxd" (OuterVolumeSpecName: "kube-api-access-5blxd") pod "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" (UID: "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1"). InnerVolumeSpecName "kube-api-access-5blxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.263806 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" (UID: "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.283919 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-config-data" (OuterVolumeSpecName: "config-data") pod "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" (UID: "1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.322214 4750 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.322254 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.322273 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5blxd\" (UniqueName: \"kubernetes.io/projected/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-kube-api-access-5blxd\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.322290 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.769046 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551381-lcv2d" event={"ID":"1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1","Type":"ContainerDied","Data":"e472f2bfd271f214009182f3ef8734a21fb766f44d363a34993784c2b02f11ba"} Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.769107 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e472f2bfd271f214009182f3ef8734a21fb766f44d363a34993784c2b02f11ba" Mar 09 19:01:05 crc kubenswrapper[4750]: I0309 19:01:05.769112 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551381-lcv2d" Mar 09 19:01:10 crc kubenswrapper[4750]: I0309 19:01:10.823530 4750 generic.go:334] "Generic (PLEG): container finished" podID="50969409-de1c-43a2-bf66-60da87e8bcb4" containerID="f8d6c4ba32da9af38c8ffca9d99d2d2b9403fc44cc0bce543821275d776796db" exitCode=0 Mar 09 19:01:10 crc kubenswrapper[4750]: I0309 19:01:10.823665 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" event={"ID":"50969409-de1c-43a2-bf66-60da87e8bcb4","Type":"ContainerDied","Data":"f8d6c4ba32da9af38c8ffca9d99d2d2b9403fc44cc0bce543821275d776796db"} Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.394081 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.484930 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-inventory\") pod \"50969409-de1c-43a2-bf66-60da87e8bcb4\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.485206 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-ssh-key-openstack-edpm-ipam\") pod \"50969409-de1c-43a2-bf66-60da87e8bcb4\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.485295 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94rhg\" (UniqueName: \"kubernetes.io/projected/50969409-de1c-43a2-bf66-60da87e8bcb4-kube-api-access-94rhg\") pod \"50969409-de1c-43a2-bf66-60da87e8bcb4\" (UID: \"50969409-de1c-43a2-bf66-60da87e8bcb4\") " Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.498234 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50969409-de1c-43a2-bf66-60da87e8bcb4-kube-api-access-94rhg" (OuterVolumeSpecName: "kube-api-access-94rhg") pod "50969409-de1c-43a2-bf66-60da87e8bcb4" (UID: "50969409-de1c-43a2-bf66-60da87e8bcb4"). InnerVolumeSpecName "kube-api-access-94rhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.531503 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-inventory" (OuterVolumeSpecName: "inventory") pod "50969409-de1c-43a2-bf66-60da87e8bcb4" (UID: "50969409-de1c-43a2-bf66-60da87e8bcb4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.539225 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "50969409-de1c-43a2-bf66-60da87e8bcb4" (UID: "50969409-de1c-43a2-bf66-60da87e8bcb4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.587902 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.587953 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94rhg\" (UniqueName: \"kubernetes.io/projected/50969409-de1c-43a2-bf66-60da87e8bcb4-kube-api-access-94rhg\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.587967 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50969409-de1c-43a2-bf66-60da87e8bcb4-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.853976 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" event={"ID":"50969409-de1c-43a2-bf66-60da87e8bcb4","Type":"ContainerDied","Data":"d53b113d4b962200cabf36ca9ae7656f661bdddc135e576dddb995ba4c50df8e"} Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.854032 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d53b113d4b962200cabf36ca9ae7656f661bdddc135e576dddb995ba4c50df8e" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.854058 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-96w2s" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.963298 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64"] Mar 09 19:01:12 crc kubenswrapper[4750]: E0309 19:01:12.963976 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" containerName="keystone-cron" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.964004 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" containerName="keystone-cron" Mar 09 19:01:12 crc kubenswrapper[4750]: E0309 19:01:12.964028 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50969409-de1c-43a2-bf66-60da87e8bcb4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.964039 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="50969409-de1c-43a2-bf66-60da87e8bcb4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.964320 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="50969409-de1c-43a2-bf66-60da87e8bcb4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.964346 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1" containerName="keystone-cron" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.965410 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.967751 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.968287 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.968447 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.968590 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:01:12 crc kubenswrapper[4750]: I0309 19:01:12.980288 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64"] Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.000584 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bncht\" (UniqueName: \"kubernetes.io/projected/56a126ce-a44f-449c-b0bc-05da61503d74-kube-api-access-bncht\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.000879 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.001022 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.102844 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.102904 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bncht\" (UniqueName: \"kubernetes.io/projected/56a126ce-a44f-449c-b0bc-05da61503d74-kube-api-access-bncht\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.102980 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.107534 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.108770 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.129367 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bncht\" (UniqueName: \"kubernetes.io/projected/56a126ce-a44f-449c-b0bc-05da61503d74-kube-api-access-bncht\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.286211 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.783128 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64"] Mar 09 19:01:13 crc kubenswrapper[4750]: I0309 19:01:13.863467 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" event={"ID":"56a126ce-a44f-449c-b0bc-05da61503d74","Type":"ContainerStarted","Data":"f5603a6443fba000ab70096b0d81d7148c9feff3c45d66863096cccddd358f50"} Mar 09 19:01:14 crc kubenswrapper[4750]: I0309 19:01:14.878920 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" event={"ID":"56a126ce-a44f-449c-b0bc-05da61503d74","Type":"ContainerStarted","Data":"bbdf2da3378ac01ed516aaaa3302f6bb5c59359bfa1da10ca61c4ef06a5251a1"} Mar 09 19:01:14 crc kubenswrapper[4750]: I0309 19:01:14.914749 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" podStartSLOduration=2.334103144 podStartE2EDuration="2.91472026s" podCreationTimestamp="2026-03-09 19:01:12 +0000 UTC" firstStartedPulling="2026-03-09 19:01:13.787915918 +0000 UTC m=+2155.130388336" lastFinishedPulling="2026-03-09 19:01:14.368533014 +0000 UTC m=+2155.711005452" observedRunningTime="2026-03-09 19:01:14.9062989 +0000 UTC m=+2156.248771348" watchObservedRunningTime="2026-03-09 19:01:14.91472026 +0000 UTC m=+2156.257192668" Mar 09 19:01:24 crc kubenswrapper[4750]: I0309 19:01:24.994498 4750 generic.go:334] "Generic (PLEG): container finished" podID="56a126ce-a44f-449c-b0bc-05da61503d74" containerID="bbdf2da3378ac01ed516aaaa3302f6bb5c59359bfa1da10ca61c4ef06a5251a1" exitCode=0 Mar 09 19:01:24 crc kubenswrapper[4750]: I0309 19:01:24.994571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" event={"ID":"56a126ce-a44f-449c-b0bc-05da61503d74","Type":"ContainerDied","Data":"bbdf2da3378ac01ed516aaaa3302f6bb5c59359bfa1da10ca61c4ef06a5251a1"} Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.505308 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.626562 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-inventory\") pod \"56a126ce-a44f-449c-b0bc-05da61503d74\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.626773 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-ssh-key-openstack-edpm-ipam\") pod \"56a126ce-a44f-449c-b0bc-05da61503d74\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.626829 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bncht\" (UniqueName: \"kubernetes.io/projected/56a126ce-a44f-449c-b0bc-05da61503d74-kube-api-access-bncht\") pod \"56a126ce-a44f-449c-b0bc-05da61503d74\" (UID: \"56a126ce-a44f-449c-b0bc-05da61503d74\") " Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.632705 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56a126ce-a44f-449c-b0bc-05da61503d74-kube-api-access-bncht" (OuterVolumeSpecName: "kube-api-access-bncht") pod "56a126ce-a44f-449c-b0bc-05da61503d74" (UID: "56a126ce-a44f-449c-b0bc-05da61503d74"). InnerVolumeSpecName "kube-api-access-bncht". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.680079 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "56a126ce-a44f-449c-b0bc-05da61503d74" (UID: "56a126ce-a44f-449c-b0bc-05da61503d74"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.685568 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-inventory" (OuterVolumeSpecName: "inventory") pod "56a126ce-a44f-449c-b0bc-05da61503d74" (UID: "56a126ce-a44f-449c-b0bc-05da61503d74"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.732091 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.732142 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/56a126ce-a44f-449c-b0bc-05da61503d74-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:26 crc kubenswrapper[4750]: I0309 19:01:26.732164 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bncht\" (UniqueName: \"kubernetes.io/projected/56a126ce-a44f-449c-b0bc-05da61503d74-kube-api-access-bncht\") on node \"crc\" DevicePath \"\"" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.017981 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" event={"ID":"56a126ce-a44f-449c-b0bc-05da61503d74","Type":"ContainerDied","Data":"f5603a6443fba000ab70096b0d81d7148c9feff3c45d66863096cccddd358f50"} Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.018071 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5603a6443fba000ab70096b0d81d7148c9feff3c45d66863096cccddd358f50" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.018070 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.123161 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n"] Mar 09 19:01:27 crc kubenswrapper[4750]: E0309 19:01:27.123811 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a126ce-a44f-449c-b0bc-05da61503d74" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.123833 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a126ce-a44f-449c-b0bc-05da61503d74" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.124115 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="56a126ce-a44f-449c-b0bc-05da61503d74" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.125089 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.134977 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.135203 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.135408 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.135515 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.135650 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.135674 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.136227 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.136331 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.143149 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n"] Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242106 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242163 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242227 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242263 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242302 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242332 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242446 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242475 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df2xm\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-kube-api-access-df2xm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242499 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242527 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242568 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242593 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242614 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.242654 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345462 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345530 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345559 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345668 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345691 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df2xm\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-kube-api-access-df2xm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345714 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345734 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345767 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345792 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345807 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345833 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345859 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345879 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.345921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.350291 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.350567 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.350925 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.350993 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.354176 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.354510 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.354517 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.356202 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.357063 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.357823 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.361846 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.362476 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.367326 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.367494 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df2xm\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-kube-api-access-df2xm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:27 crc kubenswrapper[4750]: I0309 19:01:27.489158 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:01:28 crc kubenswrapper[4750]: I0309 19:01:28.104400 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n"] Mar 09 19:01:29 crc kubenswrapper[4750]: I0309 19:01:29.044340 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" event={"ID":"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca","Type":"ContainerStarted","Data":"8dfb4aab562d60fab7975c44b2c4e90081e047453f9414183261e9ed524876c8"} Mar 09 19:01:29 crc kubenswrapper[4750]: I0309 19:01:29.044742 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" event={"ID":"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca","Type":"ContainerStarted","Data":"3f9920c46f893d84b4fae9bf471d6bb4e5bf69789e4fb8078db49483387712bc"} Mar 09 19:01:29 crc kubenswrapper[4750]: I0309 19:01:29.079781 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" podStartSLOduration=1.601392715 podStartE2EDuration="2.0797476s" podCreationTimestamp="2026-03-09 19:01:27 +0000 UTC" firstStartedPulling="2026-03-09 19:01:28.120938112 +0000 UTC m=+2169.463410510" lastFinishedPulling="2026-03-09 19:01:28.599292977 +0000 UTC m=+2169.941765395" observedRunningTime="2026-03-09 19:01:29.064796371 +0000 UTC m=+2170.407268809" watchObservedRunningTime="2026-03-09 19:01:29.0797476 +0000 UTC m=+2170.422220028" Mar 09 19:01:44 crc kubenswrapper[4750]: I0309 19:01:44.951531 4750 scope.go:117] "RemoveContainer" containerID="6d72aa0e053da98f908c0efe2d9060f98662621d865cc3084b9e3c0451bc1878" Mar 09 19:01:51 crc kubenswrapper[4750]: I0309 19:01:51.744149 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:01:51 crc kubenswrapper[4750]: I0309 19:01:51.744965 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.172065 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551382-798hk"] Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.175441 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551382-798hk" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.180050 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.180496 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.180761 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.184529 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551382-798hk"] Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.342116 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kfqm\" (UniqueName: \"kubernetes.io/projected/4a6d330b-4544-48cf-a6a3-501aed2cc800-kube-api-access-7kfqm\") pod \"auto-csr-approver-29551382-798hk\" (UID: \"4a6d330b-4544-48cf-a6a3-501aed2cc800\") " pod="openshift-infra/auto-csr-approver-29551382-798hk" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.444752 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kfqm\" (UniqueName: \"kubernetes.io/projected/4a6d330b-4544-48cf-a6a3-501aed2cc800-kube-api-access-7kfqm\") pod \"auto-csr-approver-29551382-798hk\" (UID: \"4a6d330b-4544-48cf-a6a3-501aed2cc800\") " pod="openshift-infra/auto-csr-approver-29551382-798hk" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.472310 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kfqm\" (UniqueName: \"kubernetes.io/projected/4a6d330b-4544-48cf-a6a3-501aed2cc800-kube-api-access-7kfqm\") pod \"auto-csr-approver-29551382-798hk\" (UID: \"4a6d330b-4544-48cf-a6a3-501aed2cc800\") " pod="openshift-infra/auto-csr-approver-29551382-798hk" Mar 09 19:02:00 crc kubenswrapper[4750]: I0309 19:02:00.509267 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551382-798hk" Mar 09 19:02:01 crc kubenswrapper[4750]: I0309 19:02:01.039231 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551382-798hk"] Mar 09 19:02:01 crc kubenswrapper[4750]: I0309 19:02:01.428816 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551382-798hk" event={"ID":"4a6d330b-4544-48cf-a6a3-501aed2cc800","Type":"ContainerStarted","Data":"e478d6983ed56ace2614c94c5ef97bc7535a5bc359d6ad86f9149ecdd6c95924"} Mar 09 19:02:02 crc kubenswrapper[4750]: I0309 19:02:02.440725 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551382-798hk" event={"ID":"4a6d330b-4544-48cf-a6a3-501aed2cc800","Type":"ContainerStarted","Data":"5ae8055273441e853eb425756a9530ec7cd8d44b0f8d53ca4e59bb29e5dff8c6"} Mar 09 19:02:02 crc kubenswrapper[4750]: I0309 19:02:02.457602 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551382-798hk" podStartSLOduration=1.494318577 podStartE2EDuration="2.457578101s" podCreationTimestamp="2026-03-09 19:02:00 +0000 UTC" firstStartedPulling="2026-03-09 19:02:01.048075574 +0000 UTC m=+2202.390547982" lastFinishedPulling="2026-03-09 19:02:02.011335078 +0000 UTC m=+2203.353807506" observedRunningTime="2026-03-09 19:02:02.455139884 +0000 UTC m=+2203.797612292" watchObservedRunningTime="2026-03-09 19:02:02.457578101 +0000 UTC m=+2203.800050499" Mar 09 19:02:03 crc kubenswrapper[4750]: I0309 19:02:03.456773 4750 generic.go:334] "Generic (PLEG): container finished" podID="4a6d330b-4544-48cf-a6a3-501aed2cc800" containerID="5ae8055273441e853eb425756a9530ec7cd8d44b0f8d53ca4e59bb29e5dff8c6" exitCode=0 Mar 09 19:02:03 crc kubenswrapper[4750]: I0309 19:02:03.456886 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551382-798hk" event={"ID":"4a6d330b-4544-48cf-a6a3-501aed2cc800","Type":"ContainerDied","Data":"5ae8055273441e853eb425756a9530ec7cd8d44b0f8d53ca4e59bb29e5dff8c6"} Mar 09 19:02:04 crc kubenswrapper[4750]: I0309 19:02:04.874832 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551382-798hk" Mar 09 19:02:04 crc kubenswrapper[4750]: I0309 19:02:04.984320 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kfqm\" (UniqueName: \"kubernetes.io/projected/4a6d330b-4544-48cf-a6a3-501aed2cc800-kube-api-access-7kfqm\") pod \"4a6d330b-4544-48cf-a6a3-501aed2cc800\" (UID: \"4a6d330b-4544-48cf-a6a3-501aed2cc800\") " Mar 09 19:02:04 crc kubenswrapper[4750]: I0309 19:02:04.994868 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a6d330b-4544-48cf-a6a3-501aed2cc800-kube-api-access-7kfqm" (OuterVolumeSpecName: "kube-api-access-7kfqm") pod "4a6d330b-4544-48cf-a6a3-501aed2cc800" (UID: "4a6d330b-4544-48cf-a6a3-501aed2cc800"). InnerVolumeSpecName "kube-api-access-7kfqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:02:05 crc kubenswrapper[4750]: I0309 19:02:05.086697 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kfqm\" (UniqueName: \"kubernetes.io/projected/4a6d330b-4544-48cf-a6a3-501aed2cc800-kube-api-access-7kfqm\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:05 crc kubenswrapper[4750]: I0309 19:02:05.476975 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551382-798hk" event={"ID":"4a6d330b-4544-48cf-a6a3-501aed2cc800","Type":"ContainerDied","Data":"e478d6983ed56ace2614c94c5ef97bc7535a5bc359d6ad86f9149ecdd6c95924"} Mar 09 19:02:05 crc kubenswrapper[4750]: I0309 19:02:05.477395 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551382-798hk" Mar 09 19:02:05 crc kubenswrapper[4750]: I0309 19:02:05.477401 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e478d6983ed56ace2614c94c5ef97bc7535a5bc359d6ad86f9149ecdd6c95924" Mar 09 19:02:05 crc kubenswrapper[4750]: I0309 19:02:05.548371 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551376-ckf8j"] Mar 09 19:02:05 crc kubenswrapper[4750]: I0309 19:02:05.561046 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551376-ckf8j"] Mar 09 19:02:07 crc kubenswrapper[4750]: I0309 19:02:07.387882 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d826dd46-ddf0-4c54-a075-f59baae9aa35" path="/var/lib/kubelet/pods/d826dd46-ddf0-4c54-a075-f59baae9aa35/volumes" Mar 09 19:02:08 crc kubenswrapper[4750]: I0309 19:02:08.535121 4750 generic.go:334] "Generic (PLEG): container finished" podID="d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" containerID="8dfb4aab562d60fab7975c44b2c4e90081e047453f9414183261e9ed524876c8" exitCode=0 Mar 09 19:02:08 crc kubenswrapper[4750]: I0309 19:02:08.535216 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" event={"ID":"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca","Type":"ContainerDied","Data":"8dfb4aab562d60fab7975c44b2c4e90081e047453f9414183261e9ed524876c8"} Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.039800 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133251 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-libvirt-combined-ca-bundle\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133412 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-telemetry-combined-ca-bundle\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133458 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133553 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-bootstrap-combined-ca-bundle\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133682 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-neutron-metadata-combined-ca-bundle\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133721 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-repo-setup-combined-ca-bundle\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133765 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133799 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133858 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ovn-combined-ca-bundle\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133921 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ssh-key-openstack-edpm-ipam\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.133978 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-inventory\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.134015 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df2xm\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-kube-api-access-df2xm\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.134134 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-nova-combined-ca-bundle\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.134201 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\" (UID: \"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca\") " Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.142158 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.142612 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.143078 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.143303 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-kube-api-access-df2xm" (OuterVolumeSpecName: "kube-api-access-df2xm") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "kube-api-access-df2xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.143434 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.143790 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.144013 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.144075 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.145486 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.146782 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.153053 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.159363 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.169360 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.172246 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-inventory" (OuterVolumeSpecName: "inventory") pod "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" (UID: "d46d43ed-f599-4bdd-bacd-3e3f8bb60fca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237060 4750 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237091 4750 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237104 4750 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237115 4750 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237126 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237135 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237144 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237177 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df2xm\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-kube-api-access-df2xm\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237187 4750 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237196 4750 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237205 4750 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237213 4750 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237221 4750 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.237232 4750 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d46d43ed-f599-4bdd-bacd-3e3f8bb60fca-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.569389 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" event={"ID":"d46d43ed-f599-4bdd-bacd-3e3f8bb60fca","Type":"ContainerDied","Data":"3f9920c46f893d84b4fae9bf471d6bb4e5bf69789e4fb8078db49483387712bc"} Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.569874 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f9920c46f893d84b4fae9bf471d6bb4e5bf69789e4fb8078db49483387712bc" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.569488 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.688180 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf"] Mar 09 19:02:10 crc kubenswrapper[4750]: E0309 19:02:10.688699 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a6d330b-4544-48cf-a6a3-501aed2cc800" containerName="oc" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.688715 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a6d330b-4544-48cf-a6a3-501aed2cc800" containerName="oc" Mar 09 19:02:10 crc kubenswrapper[4750]: E0309 19:02:10.688739 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.688748 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.688939 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d46d43ed-f599-4bdd-bacd-3e3f8bb60fca" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.688968 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a6d330b-4544-48cf-a6a3-501aed2cc800" containerName="oc" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.689779 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.692608 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.698396 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf"] Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.703785 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.704021 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.704159 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.704697 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.748153 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/39d323a7-f34e-404f-ae88-9a3322950f85-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.748213 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk7s5\" (UniqueName: \"kubernetes.io/projected/39d323a7-f34e-404f-ae88-9a3322950f85-kube-api-access-wk7s5\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.748254 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.748296 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.748366 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.850178 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.850310 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/39d323a7-f34e-404f-ae88-9a3322950f85-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.850339 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk7s5\" (UniqueName: \"kubernetes.io/projected/39d323a7-f34e-404f-ae88-9a3322950f85-kube-api-access-wk7s5\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.850379 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.850425 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.851391 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/39d323a7-f34e-404f-ae88-9a3322950f85-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.855719 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.855767 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.860108 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:10 crc kubenswrapper[4750]: I0309 19:02:10.870348 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk7s5\" (UniqueName: \"kubernetes.io/projected/39d323a7-f34e-404f-ae88-9a3322950f85-kube-api-access-wk7s5\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6l9mf\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:11 crc kubenswrapper[4750]: I0309 19:02:11.039186 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:02:11 crc kubenswrapper[4750]: I0309 19:02:11.576387 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf"] Mar 09 19:02:12 crc kubenswrapper[4750]: I0309 19:02:12.600748 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" event={"ID":"39d323a7-f34e-404f-ae88-9a3322950f85","Type":"ContainerStarted","Data":"3d0323463ae8e3d241da4f505b101e77c92d3c31f551c6a2d815ab5514b3a670"} Mar 09 19:02:12 crc kubenswrapper[4750]: I0309 19:02:12.601369 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" event={"ID":"39d323a7-f34e-404f-ae88-9a3322950f85","Type":"ContainerStarted","Data":"d91c3a83da09eb7069022ac17be6a780e1078e78add4163db4f99cb1de8679c6"} Mar 09 19:02:12 crc kubenswrapper[4750]: I0309 19:02:12.625558 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" podStartSLOduration=2.190850821 podStartE2EDuration="2.625536637s" podCreationTimestamp="2026-03-09 19:02:10 +0000 UTC" firstStartedPulling="2026-03-09 19:02:11.581871234 +0000 UTC m=+2212.924343632" lastFinishedPulling="2026-03-09 19:02:12.01655705 +0000 UTC m=+2213.359029448" observedRunningTime="2026-03-09 19:02:12.620563781 +0000 UTC m=+2213.963036179" watchObservedRunningTime="2026-03-09 19:02:12.625536637 +0000 UTC m=+2213.968009045" Mar 09 19:02:21 crc kubenswrapper[4750]: I0309 19:02:21.744082 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:02:21 crc kubenswrapper[4750]: I0309 19:02:21.744972 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:02:45 crc kubenswrapper[4750]: I0309 19:02:45.043161 4750 scope.go:117] "RemoveContainer" containerID="53c05d265185f8b21273e90b0774de4a8db394cf52babbb8137ebbeaeffb1b1b" Mar 09 19:02:51 crc kubenswrapper[4750]: I0309 19:02:51.743761 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:02:51 crc kubenswrapper[4750]: I0309 19:02:51.745061 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:02:51 crc kubenswrapper[4750]: I0309 19:02:51.745162 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:02:51 crc kubenswrapper[4750]: I0309 19:02:51.746616 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9744b7c86b0d0903700b40068ae49891b7ad365bb53b7b535d0120e11bb79ec7"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:02:51 crc kubenswrapper[4750]: I0309 19:02:51.746787 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://9744b7c86b0d0903700b40068ae49891b7ad365bb53b7b535d0120e11bb79ec7" gracePeriod=600 Mar 09 19:02:52 crc kubenswrapper[4750]: I0309 19:02:52.079230 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="9744b7c86b0d0903700b40068ae49891b7ad365bb53b7b535d0120e11bb79ec7" exitCode=0 Mar 09 19:02:52 crc kubenswrapper[4750]: I0309 19:02:52.079292 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"9744b7c86b0d0903700b40068ae49891b7ad365bb53b7b535d0120e11bb79ec7"} Mar 09 19:02:52 crc kubenswrapper[4750]: I0309 19:02:52.079684 4750 scope.go:117] "RemoveContainer" containerID="617d6b985507c104cb3a2f95dd1608759a3133d7565ef07c84ce9f650acf28f1" Mar 09 19:02:53 crc kubenswrapper[4750]: I0309 19:02:53.097496 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db"} Mar 09 19:03:17 crc kubenswrapper[4750]: I0309 19:03:17.911740 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rz4tq"] Mar 09 19:03:17 crc kubenswrapper[4750]: I0309 19:03:17.915897 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:17 crc kubenswrapper[4750]: I0309 19:03:17.927148 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rz4tq"] Mar 09 19:03:17 crc kubenswrapper[4750]: I0309 19:03:17.946151 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2prf\" (UniqueName: \"kubernetes.io/projected/c32787f9-abef-430d-8860-b27f40363bc9-kube-api-access-z2prf\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:17 crc kubenswrapper[4750]: I0309 19:03:17.946240 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-catalog-content\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:17 crc kubenswrapper[4750]: I0309 19:03:17.946308 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-utilities\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.048427 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-utilities\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.048515 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2prf\" (UniqueName: \"kubernetes.io/projected/c32787f9-abef-430d-8860-b27f40363bc9-kube-api-access-z2prf\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.048596 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-catalog-content\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.048993 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-utilities\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.049012 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-catalog-content\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.072299 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2prf\" (UniqueName: \"kubernetes.io/projected/c32787f9-abef-430d-8860-b27f40363bc9-kube-api-access-z2prf\") pod \"certified-operators-rz4tq\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.250545 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:18 crc kubenswrapper[4750]: I0309 19:03:18.757837 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rz4tq"] Mar 09 19:03:19 crc kubenswrapper[4750]: I0309 19:03:19.452120 4750 generic.go:334] "Generic (PLEG): container finished" podID="c32787f9-abef-430d-8860-b27f40363bc9" containerID="7b5c76dd57685b3c8123c73c8585cc70b1702090f4fd1a86012fd80c0116017d" exitCode=0 Mar 09 19:03:19 crc kubenswrapper[4750]: I0309 19:03:19.452229 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4tq" event={"ID":"c32787f9-abef-430d-8860-b27f40363bc9","Type":"ContainerDied","Data":"7b5c76dd57685b3c8123c73c8585cc70b1702090f4fd1a86012fd80c0116017d"} Mar 09 19:03:19 crc kubenswrapper[4750]: I0309 19:03:19.453433 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4tq" event={"ID":"c32787f9-abef-430d-8860-b27f40363bc9","Type":"ContainerStarted","Data":"fc682976c7c5560704610d448ea67a33afa1cc93ecf1b04e36bb6e833fde0bc8"} Mar 09 19:03:19 crc kubenswrapper[4750]: I0309 19:03:19.454489 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:03:20 crc kubenswrapper[4750]: I0309 19:03:20.464062 4750 generic.go:334] "Generic (PLEG): container finished" podID="39d323a7-f34e-404f-ae88-9a3322950f85" containerID="3d0323463ae8e3d241da4f505b101e77c92d3c31f551c6a2d815ab5514b3a670" exitCode=0 Mar 09 19:03:20 crc kubenswrapper[4750]: I0309 19:03:20.464117 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" event={"ID":"39d323a7-f34e-404f-ae88-9a3322950f85","Type":"ContainerDied","Data":"3d0323463ae8e3d241da4f505b101e77c92d3c31f551c6a2d815ab5514b3a670"} Mar 09 19:03:21 crc kubenswrapper[4750]: I0309 19:03:21.950706 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.033084 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-inventory\") pod \"39d323a7-f34e-404f-ae88-9a3322950f85\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.033504 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ovn-combined-ca-bundle\") pod \"39d323a7-f34e-404f-ae88-9a3322950f85\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.033616 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/39d323a7-f34e-404f-ae88-9a3322950f85-ovncontroller-config-0\") pod \"39d323a7-f34e-404f-ae88-9a3322950f85\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.033805 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ssh-key-openstack-edpm-ipam\") pod \"39d323a7-f34e-404f-ae88-9a3322950f85\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.034005 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk7s5\" (UniqueName: \"kubernetes.io/projected/39d323a7-f34e-404f-ae88-9a3322950f85-kube-api-access-wk7s5\") pod \"39d323a7-f34e-404f-ae88-9a3322950f85\" (UID: \"39d323a7-f34e-404f-ae88-9a3322950f85\") " Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.057416 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39d323a7-f34e-404f-ae88-9a3322950f85-kube-api-access-wk7s5" (OuterVolumeSpecName: "kube-api-access-wk7s5") pod "39d323a7-f34e-404f-ae88-9a3322950f85" (UID: "39d323a7-f34e-404f-ae88-9a3322950f85"). InnerVolumeSpecName "kube-api-access-wk7s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.057596 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "39d323a7-f34e-404f-ae88-9a3322950f85" (UID: "39d323a7-f34e-404f-ae88-9a3322950f85"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.069814 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39d323a7-f34e-404f-ae88-9a3322950f85-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "39d323a7-f34e-404f-ae88-9a3322950f85" (UID: "39d323a7-f34e-404f-ae88-9a3322950f85"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.077815 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-inventory" (OuterVolumeSpecName: "inventory") pod "39d323a7-f34e-404f-ae88-9a3322950f85" (UID: "39d323a7-f34e-404f-ae88-9a3322950f85"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.080497 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "39d323a7-f34e-404f-ae88-9a3322950f85" (UID: "39d323a7-f34e-404f-ae88-9a3322950f85"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.136347 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.136379 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.136389 4750 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/39d323a7-f34e-404f-ae88-9a3322950f85-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.136397 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/39d323a7-f34e-404f-ae88-9a3322950f85-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.136406 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk7s5\" (UniqueName: \"kubernetes.io/projected/39d323a7-f34e-404f-ae88-9a3322950f85-kube-api-access-wk7s5\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.484205 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" event={"ID":"39d323a7-f34e-404f-ae88-9a3322950f85","Type":"ContainerDied","Data":"d91c3a83da09eb7069022ac17be6a780e1078e78add4163db4f99cb1de8679c6"} Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.484491 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d91c3a83da09eb7069022ac17be6a780e1078e78add4163db4f99cb1de8679c6" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.484453 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6l9mf" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.619422 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn"] Mar 09 19:03:22 crc kubenswrapper[4750]: E0309 19:03:22.619996 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d323a7-f34e-404f-ae88-9a3322950f85" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.620018 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d323a7-f34e-404f-ae88-9a3322950f85" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.620250 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d323a7-f34e-404f-ae88-9a3322950f85" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.620997 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.632338 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.632456 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.632487 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.632643 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.632852 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.633461 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.643869 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.643932 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.643998 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.644115 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb78h\" (UniqueName: \"kubernetes.io/projected/20df7645-4b55-487a-85c1-cb23099b9527-kube-api-access-sb78h\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.644159 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.644265 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.645877 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn"] Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.745720 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb78h\" (UniqueName: \"kubernetes.io/projected/20df7645-4b55-487a-85c1-cb23099b9527-kube-api-access-sb78h\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.745796 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.745843 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.745883 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.745907 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.745941 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.751047 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.751123 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.752390 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.752767 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.753764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.768163 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb78h\" (UniqueName: \"kubernetes.io/projected/20df7645-4b55-487a-85c1-cb23099b9527-kube-api-access-sb78h\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:22 crc kubenswrapper[4750]: I0309 19:03:22.943063 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:03:25 crc kubenswrapper[4750]: I0309 19:03:25.526464 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn"] Mar 09 19:03:25 crc kubenswrapper[4750]: I0309 19:03:25.545251 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" event={"ID":"20df7645-4b55-487a-85c1-cb23099b9527","Type":"ContainerStarted","Data":"859311a3557b511de8bcfd0deea39ce630cdfa1c04b0dead7b69aac7fcca9683"} Mar 09 19:03:25 crc kubenswrapper[4750]: I0309 19:03:25.547657 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4tq" event={"ID":"c32787f9-abef-430d-8860-b27f40363bc9","Type":"ContainerStarted","Data":"0194c2b57a19b4324297353a92635ad6352cc5df89591df85f62ca169ef773b1"} Mar 09 19:03:26 crc kubenswrapper[4750]: I0309 19:03:26.562205 4750 generic.go:334] "Generic (PLEG): container finished" podID="c32787f9-abef-430d-8860-b27f40363bc9" containerID="0194c2b57a19b4324297353a92635ad6352cc5df89591df85f62ca169ef773b1" exitCode=0 Mar 09 19:03:26 crc kubenswrapper[4750]: I0309 19:03:26.562285 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4tq" event={"ID":"c32787f9-abef-430d-8860-b27f40363bc9","Type":"ContainerDied","Data":"0194c2b57a19b4324297353a92635ad6352cc5df89591df85f62ca169ef773b1"} Mar 09 19:03:26 crc kubenswrapper[4750]: I0309 19:03:26.565472 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" event={"ID":"20df7645-4b55-487a-85c1-cb23099b9527","Type":"ContainerStarted","Data":"03515fc35e73f8ed4923a1aeba38481883bd6566747d9ad4227e3007e1a0a5f5"} Mar 09 19:03:26 crc kubenswrapper[4750]: I0309 19:03:26.609026 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" podStartSLOduration=4.062858703 podStartE2EDuration="4.608999537s" podCreationTimestamp="2026-03-09 19:03:22 +0000 UTC" firstStartedPulling="2026-03-09 19:03:25.532866588 +0000 UTC m=+2286.875338986" lastFinishedPulling="2026-03-09 19:03:26.079007422 +0000 UTC m=+2287.421479820" observedRunningTime="2026-03-09 19:03:26.603423095 +0000 UTC m=+2287.945895513" watchObservedRunningTime="2026-03-09 19:03:26.608999537 +0000 UTC m=+2287.951471975" Mar 09 19:03:27 crc kubenswrapper[4750]: I0309 19:03:27.594385 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4tq" event={"ID":"c32787f9-abef-430d-8860-b27f40363bc9","Type":"ContainerStarted","Data":"f613f84965c8c9ebc5ebaa0bc26191f9e1757e8e49807833c1c926f3c4fb9f31"} Mar 09 19:03:27 crc kubenswrapper[4750]: I0309 19:03:27.627786 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rz4tq" podStartSLOduration=3.034910648 podStartE2EDuration="10.627768663s" podCreationTimestamp="2026-03-09 19:03:17 +0000 UTC" firstStartedPulling="2026-03-09 19:03:19.45416282 +0000 UTC m=+2280.796635228" lastFinishedPulling="2026-03-09 19:03:27.047020845 +0000 UTC m=+2288.389493243" observedRunningTime="2026-03-09 19:03:27.619202749 +0000 UTC m=+2288.961675157" watchObservedRunningTime="2026-03-09 19:03:27.627768663 +0000 UTC m=+2288.970241061" Mar 09 19:03:28 crc kubenswrapper[4750]: I0309 19:03:28.251267 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:28 crc kubenswrapper[4750]: I0309 19:03:28.251744 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:29 crc kubenswrapper[4750]: I0309 19:03:29.337491 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rz4tq" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="registry-server" probeResult="failure" output=< Mar 09 19:03:29 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:03:29 crc kubenswrapper[4750]: > Mar 09 19:03:38 crc kubenswrapper[4750]: I0309 19:03:38.343029 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:38 crc kubenswrapper[4750]: I0309 19:03:38.424730 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:03:38 crc kubenswrapper[4750]: I0309 19:03:38.520234 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rz4tq"] Mar 09 19:03:38 crc kubenswrapper[4750]: I0309 19:03:38.607091 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wtbl2"] Mar 09 19:03:38 crc kubenswrapper[4750]: I0309 19:03:38.607323 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wtbl2" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="registry-server" containerID="cri-o://0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0" gracePeriod=2 Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.689022 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.729971 4750 generic.go:334] "Generic (PLEG): container finished" podID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerID="0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0" exitCode=0 Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.730046 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtbl2" event={"ID":"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83","Type":"ContainerDied","Data":"0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0"} Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.730098 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtbl2" event={"ID":"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83","Type":"ContainerDied","Data":"01bd8be7c16c3a58b7410c558ad69a65fd9a10329690ce54b8c7ac02de4e4837"} Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.730120 4750 scope.go:117] "RemoveContainer" containerID="0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.730246 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtbl2" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.755616 4750 scope.go:117] "RemoveContainer" containerID="66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.780676 4750 scope.go:117] "RemoveContainer" containerID="31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.817465 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt87m\" (UniqueName: \"kubernetes.io/projected/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-kube-api-access-qt87m\") pod \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.817658 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-catalog-content\") pod \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.817741 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-utilities\") pod \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\" (UID: \"1ee24421-1fb6-4417-8a9d-54eb3fcf8b83\") " Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.819854 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-utilities" (OuterVolumeSpecName: "utilities") pod "1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" (UID: "1ee24421-1fb6-4417-8a9d-54eb3fcf8b83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.825913 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-kube-api-access-qt87m" (OuterVolumeSpecName: "kube-api-access-qt87m") pod "1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" (UID: "1ee24421-1fb6-4417-8a9d-54eb3fcf8b83"). InnerVolumeSpecName "kube-api-access-qt87m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.832760 4750 scope.go:117] "RemoveContainer" containerID="0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0" Mar 09 19:03:39 crc kubenswrapper[4750]: E0309 19:03:39.833620 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0\": container with ID starting with 0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0 not found: ID does not exist" containerID="0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.833671 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0"} err="failed to get container status \"0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0\": rpc error: code = NotFound desc = could not find container \"0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0\": container with ID starting with 0c0e80d7afa364dd9bd769eb11cac748ea08f4bb6554ff0044898f910d6890f0 not found: ID does not exist" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.833692 4750 scope.go:117] "RemoveContainer" containerID="66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834" Mar 09 19:03:39 crc kubenswrapper[4750]: E0309 19:03:39.834020 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834\": container with ID starting with 66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834 not found: ID does not exist" containerID="66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.834043 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834"} err="failed to get container status \"66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834\": rpc error: code = NotFound desc = could not find container \"66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834\": container with ID starting with 66945fe2e5dd5e475349346b82b5f68dff8088de10ef0f2c1624522732b52834 not found: ID does not exist" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.834057 4750 scope.go:117] "RemoveContainer" containerID="31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c" Mar 09 19:03:39 crc kubenswrapper[4750]: E0309 19:03:39.834284 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c\": container with ID starting with 31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c not found: ID does not exist" containerID="31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.834302 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c"} err="failed to get container status \"31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c\": rpc error: code = NotFound desc = could not find container \"31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c\": container with ID starting with 31d9d0c384e8f535a4f052008fa7ca01caf1a7d7c7fbdcd188bb9da20ff6322c not found: ID does not exist" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.913352 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" (UID: "1ee24421-1fb6-4417-8a9d-54eb3fcf8b83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.920822 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.920850 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:39 crc kubenswrapper[4750]: I0309 19:03:39.920864 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt87m\" (UniqueName: \"kubernetes.io/projected/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83-kube-api-access-qt87m\") on node \"crc\" DevicePath \"\"" Mar 09 19:03:40 crc kubenswrapper[4750]: I0309 19:03:40.071301 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wtbl2"] Mar 09 19:03:40 crc kubenswrapper[4750]: I0309 19:03:40.079107 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wtbl2"] Mar 09 19:03:41 crc kubenswrapper[4750]: I0309 19:03:41.386254 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" path="/var/lib/kubelet/pods/1ee24421-1fb6-4417-8a9d-54eb3fcf8b83/volumes" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.877770 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r94x8"] Mar 09 19:03:46 crc kubenswrapper[4750]: E0309 19:03:46.879331 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="extract-content" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.879357 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="extract-content" Mar 09 19:03:46 crc kubenswrapper[4750]: E0309 19:03:46.879420 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="registry-server" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.879434 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="registry-server" Mar 09 19:03:46 crc kubenswrapper[4750]: E0309 19:03:46.879459 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="extract-utilities" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.879472 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="extract-utilities" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.879869 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ee24421-1fb6-4417-8a9d-54eb3fcf8b83" containerName="registry-server" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.882382 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.918174 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r94x8"] Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.980097 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-catalog-content\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.980140 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-utilities\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:46 crc kubenswrapper[4750]: I0309 19:03:46.980179 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7b94\" (UniqueName: \"kubernetes.io/projected/1307eee3-8782-4211-bdba-c4c1debbf759-kube-api-access-w7b94\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.081727 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-catalog-content\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.081771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-utilities\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.081838 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7b94\" (UniqueName: \"kubernetes.io/projected/1307eee3-8782-4211-bdba-c4c1debbf759-kube-api-access-w7b94\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.082589 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-catalog-content\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.082751 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-utilities\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.111768 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7b94\" (UniqueName: \"kubernetes.io/projected/1307eee3-8782-4211-bdba-c4c1debbf759-kube-api-access-w7b94\") pod \"redhat-marketplace-r94x8\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.225456 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.717615 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r94x8"] Mar 09 19:03:47 crc kubenswrapper[4750]: W0309 19:03:47.720480 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1307eee3_8782_4211_bdba_c4c1debbf759.slice/crio-f7ebde0788ebb858dc3c4159ca89c06c592982be8969676beb22601048c6efc1 WatchSource:0}: Error finding container f7ebde0788ebb858dc3c4159ca89c06c592982be8969676beb22601048c6efc1: Status 404 returned error can't find the container with id f7ebde0788ebb858dc3c4159ca89c06c592982be8969676beb22601048c6efc1 Mar 09 19:03:47 crc kubenswrapper[4750]: I0309 19:03:47.854731 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r94x8" event={"ID":"1307eee3-8782-4211-bdba-c4c1debbf759","Type":"ContainerStarted","Data":"f7ebde0788ebb858dc3c4159ca89c06c592982be8969676beb22601048c6efc1"} Mar 09 19:03:48 crc kubenswrapper[4750]: I0309 19:03:48.870014 4750 generic.go:334] "Generic (PLEG): container finished" podID="1307eee3-8782-4211-bdba-c4c1debbf759" containerID="f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03" exitCode=0 Mar 09 19:03:48 crc kubenswrapper[4750]: I0309 19:03:48.870076 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r94x8" event={"ID":"1307eee3-8782-4211-bdba-c4c1debbf759","Type":"ContainerDied","Data":"f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03"} Mar 09 19:03:49 crc kubenswrapper[4750]: I0309 19:03:49.882266 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r94x8" event={"ID":"1307eee3-8782-4211-bdba-c4c1debbf759","Type":"ContainerStarted","Data":"82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10"} Mar 09 19:03:51 crc kubenswrapper[4750]: I0309 19:03:51.901942 4750 generic.go:334] "Generic (PLEG): container finished" podID="1307eee3-8782-4211-bdba-c4c1debbf759" containerID="82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10" exitCode=0 Mar 09 19:03:51 crc kubenswrapper[4750]: I0309 19:03:51.902027 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r94x8" event={"ID":"1307eee3-8782-4211-bdba-c4c1debbf759","Type":"ContainerDied","Data":"82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10"} Mar 09 19:03:52 crc kubenswrapper[4750]: I0309 19:03:52.919129 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r94x8" event={"ID":"1307eee3-8782-4211-bdba-c4c1debbf759","Type":"ContainerStarted","Data":"ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730"} Mar 09 19:03:52 crc kubenswrapper[4750]: I0309 19:03:52.946959 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r94x8" podStartSLOduration=3.360020881 podStartE2EDuration="6.946943299s" podCreationTimestamp="2026-03-09 19:03:46 +0000 UTC" firstStartedPulling="2026-03-09 19:03:48.873341876 +0000 UTC m=+2310.215814284" lastFinishedPulling="2026-03-09 19:03:52.460264264 +0000 UTC m=+2313.802736702" observedRunningTime="2026-03-09 19:03:52.937514991 +0000 UTC m=+2314.279987389" watchObservedRunningTime="2026-03-09 19:03:52.946943299 +0000 UTC m=+2314.289415697" Mar 09 19:03:57 crc kubenswrapper[4750]: I0309 19:03:57.225900 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:57 crc kubenswrapper[4750]: I0309 19:03:57.227769 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:03:58 crc kubenswrapper[4750]: I0309 19:03:58.324225 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-r94x8" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="registry-server" probeResult="failure" output=< Mar 09 19:03:58 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:03:58 crc kubenswrapper[4750]: > Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.154769 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551384-n8pxf"] Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.158471 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551384-n8pxf" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.166469 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.166479 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.166992 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.189183 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551384-n8pxf"] Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.299033 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwv8l\" (UniqueName: \"kubernetes.io/projected/010e3beb-8423-4ab0-b4b4-5f3344e594ed-kube-api-access-kwv8l\") pod \"auto-csr-approver-29551384-n8pxf\" (UID: \"010e3beb-8423-4ab0-b4b4-5f3344e594ed\") " pod="openshift-infra/auto-csr-approver-29551384-n8pxf" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.401415 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwv8l\" (UniqueName: \"kubernetes.io/projected/010e3beb-8423-4ab0-b4b4-5f3344e594ed-kube-api-access-kwv8l\") pod \"auto-csr-approver-29551384-n8pxf\" (UID: \"010e3beb-8423-4ab0-b4b4-5f3344e594ed\") " pod="openshift-infra/auto-csr-approver-29551384-n8pxf" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.434603 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwv8l\" (UniqueName: \"kubernetes.io/projected/010e3beb-8423-4ab0-b4b4-5f3344e594ed-kube-api-access-kwv8l\") pod \"auto-csr-approver-29551384-n8pxf\" (UID: \"010e3beb-8423-4ab0-b4b4-5f3344e594ed\") " pod="openshift-infra/auto-csr-approver-29551384-n8pxf" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.493890 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551384-n8pxf" Mar 09 19:04:00 crc kubenswrapper[4750]: I0309 19:04:00.786212 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551384-n8pxf"] Mar 09 19:04:01 crc kubenswrapper[4750]: I0309 19:04:01.004061 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551384-n8pxf" event={"ID":"010e3beb-8423-4ab0-b4b4-5f3344e594ed","Type":"ContainerStarted","Data":"8207b43f27f1e0f268d04b3f1cadbddc817e5c60667bba72bee180da7341c52a"} Mar 09 19:04:03 crc kubenswrapper[4750]: I0309 19:04:03.047880 4750 generic.go:334] "Generic (PLEG): container finished" podID="010e3beb-8423-4ab0-b4b4-5f3344e594ed" containerID="18852841a2a023c9eb2f2433a02cfe1c2cca08ca9365dd8b34296418cafa21f0" exitCode=0 Mar 09 19:04:03 crc kubenswrapper[4750]: I0309 19:04:03.047982 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551384-n8pxf" event={"ID":"010e3beb-8423-4ab0-b4b4-5f3344e594ed","Type":"ContainerDied","Data":"18852841a2a023c9eb2f2433a02cfe1c2cca08ca9365dd8b34296418cafa21f0"} Mar 09 19:04:04 crc kubenswrapper[4750]: I0309 19:04:04.412027 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551384-n8pxf" Mar 09 19:04:04 crc kubenswrapper[4750]: I0309 19:04:04.592274 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwv8l\" (UniqueName: \"kubernetes.io/projected/010e3beb-8423-4ab0-b4b4-5f3344e594ed-kube-api-access-kwv8l\") pod \"010e3beb-8423-4ab0-b4b4-5f3344e594ed\" (UID: \"010e3beb-8423-4ab0-b4b4-5f3344e594ed\") " Mar 09 19:04:04 crc kubenswrapper[4750]: I0309 19:04:04.598866 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/010e3beb-8423-4ab0-b4b4-5f3344e594ed-kube-api-access-kwv8l" (OuterVolumeSpecName: "kube-api-access-kwv8l") pod "010e3beb-8423-4ab0-b4b4-5f3344e594ed" (UID: "010e3beb-8423-4ab0-b4b4-5f3344e594ed"). InnerVolumeSpecName "kube-api-access-kwv8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:04:04 crc kubenswrapper[4750]: I0309 19:04:04.695695 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwv8l\" (UniqueName: \"kubernetes.io/projected/010e3beb-8423-4ab0-b4b4-5f3344e594ed-kube-api-access-kwv8l\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:05 crc kubenswrapper[4750]: I0309 19:04:05.074253 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551384-n8pxf" event={"ID":"010e3beb-8423-4ab0-b4b4-5f3344e594ed","Type":"ContainerDied","Data":"8207b43f27f1e0f268d04b3f1cadbddc817e5c60667bba72bee180da7341c52a"} Mar 09 19:04:05 crc kubenswrapper[4750]: I0309 19:04:05.074316 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8207b43f27f1e0f268d04b3f1cadbddc817e5c60667bba72bee180da7341c52a" Mar 09 19:04:05 crc kubenswrapper[4750]: I0309 19:04:05.074339 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551384-n8pxf" Mar 09 19:04:05 crc kubenswrapper[4750]: I0309 19:04:05.495692 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551378-tm6ft"] Mar 09 19:04:05 crc kubenswrapper[4750]: I0309 19:04:05.506884 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551378-tm6ft"] Mar 09 19:04:07 crc kubenswrapper[4750]: I0309 19:04:07.310847 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:04:07 crc kubenswrapper[4750]: I0309 19:04:07.387321 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="004b9dfd-4abd-449e-87f1-a665e958db40" path="/var/lib/kubelet/pods/004b9dfd-4abd-449e-87f1-a665e958db40/volumes" Mar 09 19:04:07 crc kubenswrapper[4750]: I0309 19:04:07.395204 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:04:07 crc kubenswrapper[4750]: I0309 19:04:07.570985 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r94x8"] Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.121125 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r94x8" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="registry-server" containerID="cri-o://ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730" gracePeriod=2 Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.603075 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.702334 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7b94\" (UniqueName: \"kubernetes.io/projected/1307eee3-8782-4211-bdba-c4c1debbf759-kube-api-access-w7b94\") pod \"1307eee3-8782-4211-bdba-c4c1debbf759\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.702463 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-utilities\") pod \"1307eee3-8782-4211-bdba-c4c1debbf759\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.702598 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-catalog-content\") pod \"1307eee3-8782-4211-bdba-c4c1debbf759\" (UID: \"1307eee3-8782-4211-bdba-c4c1debbf759\") " Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.703138 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-utilities" (OuterVolumeSpecName: "utilities") pod "1307eee3-8782-4211-bdba-c4c1debbf759" (UID: "1307eee3-8782-4211-bdba-c4c1debbf759"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.703263 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.718507 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1307eee3-8782-4211-bdba-c4c1debbf759-kube-api-access-w7b94" (OuterVolumeSpecName: "kube-api-access-w7b94") pod "1307eee3-8782-4211-bdba-c4c1debbf759" (UID: "1307eee3-8782-4211-bdba-c4c1debbf759"). InnerVolumeSpecName "kube-api-access-w7b94". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.734025 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1307eee3-8782-4211-bdba-c4c1debbf759" (UID: "1307eee3-8782-4211-bdba-c4c1debbf759"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.805485 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1307eee3-8782-4211-bdba-c4c1debbf759-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:09 crc kubenswrapper[4750]: I0309 19:04:09.805532 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7b94\" (UniqueName: \"kubernetes.io/projected/1307eee3-8782-4211-bdba-c4c1debbf759-kube-api-access-w7b94\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.134896 4750 generic.go:334] "Generic (PLEG): container finished" podID="1307eee3-8782-4211-bdba-c4c1debbf759" containerID="ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730" exitCode=0 Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.134957 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r94x8" event={"ID":"1307eee3-8782-4211-bdba-c4c1debbf759","Type":"ContainerDied","Data":"ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730"} Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.135019 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r94x8" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.135048 4750 scope.go:117] "RemoveContainer" containerID="ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.135030 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r94x8" event={"ID":"1307eee3-8782-4211-bdba-c4c1debbf759","Type":"ContainerDied","Data":"f7ebde0788ebb858dc3c4159ca89c06c592982be8969676beb22601048c6efc1"} Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.168109 4750 scope.go:117] "RemoveContainer" containerID="82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.202677 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r94x8"] Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.214383 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r94x8"] Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.215219 4750 scope.go:117] "RemoveContainer" containerID="f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.251216 4750 scope.go:117] "RemoveContainer" containerID="ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730" Mar 09 19:04:10 crc kubenswrapper[4750]: E0309 19:04:10.251684 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730\": container with ID starting with ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730 not found: ID does not exist" containerID="ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.251749 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730"} err="failed to get container status \"ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730\": rpc error: code = NotFound desc = could not find container \"ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730\": container with ID starting with ad21964eed424cd8f4e2c4cc8064fd6eed4f1d2c998bb8002aa11d65da065730 not found: ID does not exist" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.251800 4750 scope.go:117] "RemoveContainer" containerID="82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10" Mar 09 19:04:10 crc kubenswrapper[4750]: E0309 19:04:10.252132 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10\": container with ID starting with 82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10 not found: ID does not exist" containerID="82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.252205 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10"} err="failed to get container status \"82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10\": rpc error: code = NotFound desc = could not find container \"82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10\": container with ID starting with 82762dfa620d3f261641c2307805c116b76ec9b6d9acb2db5ee16f8c2fe2dd10 not found: ID does not exist" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.252278 4750 scope.go:117] "RemoveContainer" containerID="f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03" Mar 09 19:04:10 crc kubenswrapper[4750]: E0309 19:04:10.252810 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03\": container with ID starting with f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03 not found: ID does not exist" containerID="f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03" Mar 09 19:04:10 crc kubenswrapper[4750]: I0309 19:04:10.252888 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03"} err="failed to get container status \"f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03\": rpc error: code = NotFound desc = could not find container \"f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03\": container with ID starting with f5adac36445a25bfd35222b7f6111ecbeac14ab671c7c4da20b17324ca21ea03 not found: ID does not exist" Mar 09 19:04:11 crc kubenswrapper[4750]: I0309 19:04:11.393853 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" path="/var/lib/kubelet/pods/1307eee3-8782-4211-bdba-c4c1debbf759/volumes" Mar 09 19:04:17 crc kubenswrapper[4750]: I0309 19:04:17.243198 4750 generic.go:334] "Generic (PLEG): container finished" podID="20df7645-4b55-487a-85c1-cb23099b9527" containerID="03515fc35e73f8ed4923a1aeba38481883bd6566747d9ad4227e3007e1a0a5f5" exitCode=0 Mar 09 19:04:17 crc kubenswrapper[4750]: I0309 19:04:17.243277 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" event={"ID":"20df7645-4b55-487a-85c1-cb23099b9527","Type":"ContainerDied","Data":"03515fc35e73f8ed4923a1aeba38481883bd6566747d9ad4227e3007e1a0a5f5"} Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.787424 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.926284 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-metadata-combined-ca-bundle\") pod \"20df7645-4b55-487a-85c1-cb23099b9527\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.926985 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-ovn-metadata-agent-neutron-config-0\") pod \"20df7645-4b55-487a-85c1-cb23099b9527\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.927326 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-inventory\") pod \"20df7645-4b55-487a-85c1-cb23099b9527\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.927459 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-nova-metadata-neutron-config-0\") pod \"20df7645-4b55-487a-85c1-cb23099b9527\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.927597 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb78h\" (UniqueName: \"kubernetes.io/projected/20df7645-4b55-487a-85c1-cb23099b9527-kube-api-access-sb78h\") pod \"20df7645-4b55-487a-85c1-cb23099b9527\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.928149 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-ssh-key-openstack-edpm-ipam\") pod \"20df7645-4b55-487a-85c1-cb23099b9527\" (UID: \"20df7645-4b55-487a-85c1-cb23099b9527\") " Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.933329 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20df7645-4b55-487a-85c1-cb23099b9527-kube-api-access-sb78h" (OuterVolumeSpecName: "kube-api-access-sb78h") pod "20df7645-4b55-487a-85c1-cb23099b9527" (UID: "20df7645-4b55-487a-85c1-cb23099b9527"). InnerVolumeSpecName "kube-api-access-sb78h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.944499 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "20df7645-4b55-487a-85c1-cb23099b9527" (UID: "20df7645-4b55-487a-85c1-cb23099b9527"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.963428 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "20df7645-4b55-487a-85c1-cb23099b9527" (UID: "20df7645-4b55-487a-85c1-cb23099b9527"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.965396 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "20df7645-4b55-487a-85c1-cb23099b9527" (UID: "20df7645-4b55-487a-85c1-cb23099b9527"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.980323 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "20df7645-4b55-487a-85c1-cb23099b9527" (UID: "20df7645-4b55-487a-85c1-cb23099b9527"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:04:18 crc kubenswrapper[4750]: I0309 19:04:18.992845 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-inventory" (OuterVolumeSpecName: "inventory") pod "20df7645-4b55-487a-85c1-cb23099b9527" (UID: "20df7645-4b55-487a-85c1-cb23099b9527"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.030872 4750 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.030913 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.030926 4750 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.030940 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb78h\" (UniqueName: \"kubernetes.io/projected/20df7645-4b55-487a-85c1-cb23099b9527-kube-api-access-sb78h\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.030952 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.030963 4750 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20df7645-4b55-487a-85c1-cb23099b9527-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.272596 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" event={"ID":"20df7645-4b55-487a-85c1-cb23099b9527","Type":"ContainerDied","Data":"859311a3557b511de8bcfd0deea39ce630cdfa1c04b0dead7b69aac7fcca9683"} Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.272644 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.272736 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="859311a3557b511de8bcfd0deea39ce630cdfa1c04b0dead7b69aac7fcca9683" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.386908 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf"] Mar 09 19:04:19 crc kubenswrapper[4750]: E0309 19:04:19.387341 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="extract-content" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387364 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="extract-content" Mar 09 19:04:19 crc kubenswrapper[4750]: E0309 19:04:19.387382 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20df7645-4b55-487a-85c1-cb23099b9527" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387397 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="20df7645-4b55-487a-85c1-cb23099b9527" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 09 19:04:19 crc kubenswrapper[4750]: E0309 19:04:19.387423 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="registry-server" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387448 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="registry-server" Mar 09 19:04:19 crc kubenswrapper[4750]: E0309 19:04:19.387472 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="extract-utilities" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387481 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="extract-utilities" Mar 09 19:04:19 crc kubenswrapper[4750]: E0309 19:04:19.387493 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="010e3beb-8423-4ab0-b4b4-5f3344e594ed" containerName="oc" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387502 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="010e3beb-8423-4ab0-b4b4-5f3344e594ed" containerName="oc" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387787 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="20df7645-4b55-487a-85c1-cb23099b9527" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387820 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="010e3beb-8423-4ab0-b4b4-5f3344e594ed" containerName="oc" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.387840 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1307eee3-8782-4211-bdba-c4c1debbf759" containerName="registry-server" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.388607 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.399678 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.399907 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.401992 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.402346 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.406666 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf"] Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.410575 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.540435 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.540719 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.540761 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.540794 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.541052 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvkbz\" (UniqueName: \"kubernetes.io/projected/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-kube-api-access-cvkbz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.643261 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.643337 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvkbz\" (UniqueName: \"kubernetes.io/projected/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-kube-api-access-cvkbz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.643466 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.643494 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.643518 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.645786 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.645950 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.650455 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.656475 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.657299 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.658923 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.661597 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.664652 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvkbz\" (UniqueName: \"kubernetes.io/projected/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-kube-api-access-cvkbz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-827lf\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.735522 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:04:19 crc kubenswrapper[4750]: I0309 19:04:19.743512 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:04:20 crc kubenswrapper[4750]: I0309 19:04:20.296256 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf"] Mar 09 19:04:20 crc kubenswrapper[4750]: I0309 19:04:20.827265 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:04:21 crc kubenswrapper[4750]: I0309 19:04:21.293732 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" event={"ID":"bcbdff51-5cc0-4011-a1ae-a260ee565ef7","Type":"ContainerStarted","Data":"f8dbf37238ea5fba19323dacaca982810dbf18387fe139a7555f003e460f9304"} Mar 09 19:04:21 crc kubenswrapper[4750]: I0309 19:04:21.293776 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" event={"ID":"bcbdff51-5cc0-4011-a1ae-a260ee565ef7","Type":"ContainerStarted","Data":"e964870eb9fa9f8152ba6cdd3b7cff69bf2af4f2cd923b15571ee11767bff1f7"} Mar 09 19:04:21 crc kubenswrapper[4750]: I0309 19:04:21.311995 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" podStartSLOduration=1.7922283449999998 podStartE2EDuration="2.311966501s" podCreationTimestamp="2026-03-09 19:04:19 +0000 UTC" firstStartedPulling="2026-03-09 19:04:20.303781214 +0000 UTC m=+2341.646253612" lastFinishedPulling="2026-03-09 19:04:20.82351935 +0000 UTC m=+2342.165991768" observedRunningTime="2026-03-09 19:04:21.307034717 +0000 UTC m=+2342.649507115" watchObservedRunningTime="2026-03-09 19:04:21.311966501 +0000 UTC m=+2342.654438929" Mar 09 19:04:45 crc kubenswrapper[4750]: I0309 19:04:45.226870 4750 scope.go:117] "RemoveContainer" containerID="cb3b6f5397daa45f40f83323b70054bd4567857753ee307e3e6ca6a1fce62660" Mar 09 19:05:21 crc kubenswrapper[4750]: I0309 19:05:21.744129 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:05:21 crc kubenswrapper[4750]: I0309 19:05:21.744744 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:05:51 crc kubenswrapper[4750]: I0309 19:05:51.744001 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:05:51 crc kubenswrapper[4750]: I0309 19:05:51.744673 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.186194 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551386-ssglf"] Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.189523 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551386-ssglf" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.196143 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.196358 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.196497 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.203745 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551386-ssglf"] Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.266477 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84vh2\" (UniqueName: \"kubernetes.io/projected/3b8c96fa-17c8-48fd-9d78-8155c97a3748-kube-api-access-84vh2\") pod \"auto-csr-approver-29551386-ssglf\" (UID: \"3b8c96fa-17c8-48fd-9d78-8155c97a3748\") " pod="openshift-infra/auto-csr-approver-29551386-ssglf" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.368412 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84vh2\" (UniqueName: \"kubernetes.io/projected/3b8c96fa-17c8-48fd-9d78-8155c97a3748-kube-api-access-84vh2\") pod \"auto-csr-approver-29551386-ssglf\" (UID: \"3b8c96fa-17c8-48fd-9d78-8155c97a3748\") " pod="openshift-infra/auto-csr-approver-29551386-ssglf" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.385949 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84vh2\" (UniqueName: \"kubernetes.io/projected/3b8c96fa-17c8-48fd-9d78-8155c97a3748-kube-api-access-84vh2\") pod \"auto-csr-approver-29551386-ssglf\" (UID: \"3b8c96fa-17c8-48fd-9d78-8155c97a3748\") " pod="openshift-infra/auto-csr-approver-29551386-ssglf" Mar 09 19:06:00 crc kubenswrapper[4750]: I0309 19:06:00.535024 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551386-ssglf" Mar 09 19:06:01 crc kubenswrapper[4750]: I0309 19:06:01.016405 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551386-ssglf"] Mar 09 19:06:01 crc kubenswrapper[4750]: I0309 19:06:01.500690 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551386-ssglf" event={"ID":"3b8c96fa-17c8-48fd-9d78-8155c97a3748","Type":"ContainerStarted","Data":"8f6ab5d722028e7f2c49c8128841d0cd70d5ce669296dc9d205a3521ae53796c"} Mar 09 19:06:03 crc kubenswrapper[4750]: I0309 19:06:03.526853 4750 generic.go:334] "Generic (PLEG): container finished" podID="3b8c96fa-17c8-48fd-9d78-8155c97a3748" containerID="6b7a52fe3fe6a8a8fa77132f544b14e8c8230aae776cc01eb98d3d7e20a546e7" exitCode=0 Mar 09 19:06:03 crc kubenswrapper[4750]: I0309 19:06:03.526958 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551386-ssglf" event={"ID":"3b8c96fa-17c8-48fd-9d78-8155c97a3748","Type":"ContainerDied","Data":"6b7a52fe3fe6a8a8fa77132f544b14e8c8230aae776cc01eb98d3d7e20a546e7"} Mar 09 19:06:05 crc kubenswrapper[4750]: I0309 19:06:05.045118 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551386-ssglf" Mar 09 19:06:05 crc kubenswrapper[4750]: I0309 19:06:05.169652 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84vh2\" (UniqueName: \"kubernetes.io/projected/3b8c96fa-17c8-48fd-9d78-8155c97a3748-kube-api-access-84vh2\") pod \"3b8c96fa-17c8-48fd-9d78-8155c97a3748\" (UID: \"3b8c96fa-17c8-48fd-9d78-8155c97a3748\") " Mar 09 19:06:05 crc kubenswrapper[4750]: I0309 19:06:05.183285 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8c96fa-17c8-48fd-9d78-8155c97a3748-kube-api-access-84vh2" (OuterVolumeSpecName: "kube-api-access-84vh2") pod "3b8c96fa-17c8-48fd-9d78-8155c97a3748" (UID: "3b8c96fa-17c8-48fd-9d78-8155c97a3748"). InnerVolumeSpecName "kube-api-access-84vh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:06:05 crc kubenswrapper[4750]: I0309 19:06:05.272115 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84vh2\" (UniqueName: \"kubernetes.io/projected/3b8c96fa-17c8-48fd-9d78-8155c97a3748-kube-api-access-84vh2\") on node \"crc\" DevicePath \"\"" Mar 09 19:06:05 crc kubenswrapper[4750]: I0309 19:06:05.548281 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551386-ssglf" event={"ID":"3b8c96fa-17c8-48fd-9d78-8155c97a3748","Type":"ContainerDied","Data":"8f6ab5d722028e7f2c49c8128841d0cd70d5ce669296dc9d205a3521ae53796c"} Mar 09 19:06:05 crc kubenswrapper[4750]: I0309 19:06:05.548332 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551386-ssglf" Mar 09 19:06:05 crc kubenswrapper[4750]: I0309 19:06:05.548336 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f6ab5d722028e7f2c49c8128841d0cd70d5ce669296dc9d205a3521ae53796c" Mar 09 19:06:06 crc kubenswrapper[4750]: I0309 19:06:06.153697 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551380-5m6ww"] Mar 09 19:06:06 crc kubenswrapper[4750]: I0309 19:06:06.165897 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551380-5m6ww"] Mar 09 19:06:07 crc kubenswrapper[4750]: I0309 19:06:07.388129 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a135e52-9627-4241-b557-a5146429cdec" path="/var/lib/kubelet/pods/2a135e52-9627-4241-b557-a5146429cdec/volumes" Mar 09 19:06:21 crc kubenswrapper[4750]: I0309 19:06:21.743593 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:06:21 crc kubenswrapper[4750]: I0309 19:06:21.744218 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:06:21 crc kubenswrapper[4750]: I0309 19:06:21.744267 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:06:21 crc kubenswrapper[4750]: I0309 19:06:21.744942 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:06:21 crc kubenswrapper[4750]: I0309 19:06:21.745014 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" gracePeriod=600 Mar 09 19:06:21 crc kubenswrapper[4750]: E0309 19:06:21.878970 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:06:22 crc kubenswrapper[4750]: I0309 19:06:22.734294 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" exitCode=0 Mar 09 19:06:22 crc kubenswrapper[4750]: I0309 19:06:22.734368 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db"} Mar 09 19:06:22 crc kubenswrapper[4750]: I0309 19:06:22.734421 4750 scope.go:117] "RemoveContainer" containerID="9744b7c86b0d0903700b40068ae49891b7ad365bb53b7b535d0120e11bb79ec7" Mar 09 19:06:22 crc kubenswrapper[4750]: I0309 19:06:22.735278 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:06:22 crc kubenswrapper[4750]: E0309 19:06:22.735875 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:06:36 crc kubenswrapper[4750]: I0309 19:06:36.373431 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:06:36 crc kubenswrapper[4750]: E0309 19:06:36.374352 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:06:45 crc kubenswrapper[4750]: I0309 19:06:45.374858 4750 scope.go:117] "RemoveContainer" containerID="83f382254d36cc8e265c3d941d6a1e857ee16c6492a8ccbf68b8f8e3d3f282f6" Mar 09 19:06:49 crc kubenswrapper[4750]: I0309 19:06:49.389877 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:06:49 crc kubenswrapper[4750]: E0309 19:06:49.390858 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:07:03 crc kubenswrapper[4750]: I0309 19:07:03.373259 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:07:03 crc kubenswrapper[4750]: E0309 19:07:03.374094 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:07:16 crc kubenswrapper[4750]: I0309 19:07:16.374280 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:07:16 crc kubenswrapper[4750]: E0309 19:07:16.375380 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:07:31 crc kubenswrapper[4750]: I0309 19:07:31.373434 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:07:31 crc kubenswrapper[4750]: E0309 19:07:31.374465 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:07:42 crc kubenswrapper[4750]: I0309 19:07:42.374449 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:07:42 crc kubenswrapper[4750]: E0309 19:07:42.375513 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:07:56 crc kubenswrapper[4750]: I0309 19:07:56.373497 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:07:56 crc kubenswrapper[4750]: E0309 19:07:56.374503 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.168605 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551388-hf8p8"] Mar 09 19:08:00 crc kubenswrapper[4750]: E0309 19:08:00.169237 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8c96fa-17c8-48fd-9d78-8155c97a3748" containerName="oc" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.169248 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8c96fa-17c8-48fd-9d78-8155c97a3748" containerName="oc" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.169441 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8c96fa-17c8-48fd-9d78-8155c97a3748" containerName="oc" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.170047 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551388-hf8p8" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.173463 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.173755 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.174734 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.189877 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551388-hf8p8"] Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.215043 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29xq5\" (UniqueName: \"kubernetes.io/projected/160d1e84-3414-4465-8ad7-d142374304aa-kube-api-access-29xq5\") pod \"auto-csr-approver-29551388-hf8p8\" (UID: \"160d1e84-3414-4465-8ad7-d142374304aa\") " pod="openshift-infra/auto-csr-approver-29551388-hf8p8" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.317079 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29xq5\" (UniqueName: \"kubernetes.io/projected/160d1e84-3414-4465-8ad7-d142374304aa-kube-api-access-29xq5\") pod \"auto-csr-approver-29551388-hf8p8\" (UID: \"160d1e84-3414-4465-8ad7-d142374304aa\") " pod="openshift-infra/auto-csr-approver-29551388-hf8p8" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.341477 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29xq5\" (UniqueName: \"kubernetes.io/projected/160d1e84-3414-4465-8ad7-d142374304aa-kube-api-access-29xq5\") pod \"auto-csr-approver-29551388-hf8p8\" (UID: \"160d1e84-3414-4465-8ad7-d142374304aa\") " pod="openshift-infra/auto-csr-approver-29551388-hf8p8" Mar 09 19:08:00 crc kubenswrapper[4750]: I0309 19:08:00.494470 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551388-hf8p8" Mar 09 19:08:01 crc kubenswrapper[4750]: I0309 19:08:01.011625 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551388-hf8p8"] Mar 09 19:08:01 crc kubenswrapper[4750]: I0309 19:08:01.900413 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551388-hf8p8" event={"ID":"160d1e84-3414-4465-8ad7-d142374304aa","Type":"ContainerStarted","Data":"10e22e02bbe42b911f180c956c92464a1c88f09c2e8363bb41887f211e0a87c1"} Mar 09 19:08:02 crc kubenswrapper[4750]: I0309 19:08:02.915965 4750 generic.go:334] "Generic (PLEG): container finished" podID="160d1e84-3414-4465-8ad7-d142374304aa" containerID="fbebdef272d543507373f4753a288dccf1a57dad80d276ede115a4b663cd465d" exitCode=0 Mar 09 19:08:02 crc kubenswrapper[4750]: I0309 19:08:02.917691 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551388-hf8p8" event={"ID":"160d1e84-3414-4465-8ad7-d142374304aa","Type":"ContainerDied","Data":"fbebdef272d543507373f4753a288dccf1a57dad80d276ede115a4b663cd465d"} Mar 09 19:08:04 crc kubenswrapper[4750]: I0309 19:08:04.316828 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551388-hf8p8" Mar 09 19:08:04 crc kubenswrapper[4750]: I0309 19:08:04.405899 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29xq5\" (UniqueName: \"kubernetes.io/projected/160d1e84-3414-4465-8ad7-d142374304aa-kube-api-access-29xq5\") pod \"160d1e84-3414-4465-8ad7-d142374304aa\" (UID: \"160d1e84-3414-4465-8ad7-d142374304aa\") " Mar 09 19:08:04 crc kubenswrapper[4750]: I0309 19:08:04.411898 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/160d1e84-3414-4465-8ad7-d142374304aa-kube-api-access-29xq5" (OuterVolumeSpecName: "kube-api-access-29xq5") pod "160d1e84-3414-4465-8ad7-d142374304aa" (UID: "160d1e84-3414-4465-8ad7-d142374304aa"). InnerVolumeSpecName "kube-api-access-29xq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:08:04 crc kubenswrapper[4750]: I0309 19:08:04.509078 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29xq5\" (UniqueName: \"kubernetes.io/projected/160d1e84-3414-4465-8ad7-d142374304aa-kube-api-access-29xq5\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:04 crc kubenswrapper[4750]: I0309 19:08:04.938563 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551388-hf8p8" event={"ID":"160d1e84-3414-4465-8ad7-d142374304aa","Type":"ContainerDied","Data":"10e22e02bbe42b911f180c956c92464a1c88f09c2e8363bb41887f211e0a87c1"} Mar 09 19:08:04 crc kubenswrapper[4750]: I0309 19:08:04.938871 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10e22e02bbe42b911f180c956c92464a1c88f09c2e8363bb41887f211e0a87c1" Mar 09 19:08:04 crc kubenswrapper[4750]: I0309 19:08:04.938697 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551388-hf8p8" Mar 09 19:08:05 crc kubenswrapper[4750]: I0309 19:08:05.430007 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551382-798hk"] Mar 09 19:08:05 crc kubenswrapper[4750]: I0309 19:08:05.431861 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551382-798hk"] Mar 09 19:08:07 crc kubenswrapper[4750]: I0309 19:08:07.396412 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a6d330b-4544-48cf-a6a3-501aed2cc800" path="/var/lib/kubelet/pods/4a6d330b-4544-48cf-a6a3-501aed2cc800/volumes" Mar 09 19:08:11 crc kubenswrapper[4750]: I0309 19:08:11.374749 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:08:11 crc kubenswrapper[4750]: E0309 19:08:11.376125 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:08:15 crc kubenswrapper[4750]: I0309 19:08:15.057178 4750 generic.go:334] "Generic (PLEG): container finished" podID="bcbdff51-5cc0-4011-a1ae-a260ee565ef7" containerID="f8dbf37238ea5fba19323dacaca982810dbf18387fe139a7555f003e460f9304" exitCode=0 Mar 09 19:08:15 crc kubenswrapper[4750]: I0309 19:08:15.057272 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" event={"ID":"bcbdff51-5cc0-4011-a1ae-a260ee565ef7","Type":"ContainerDied","Data":"f8dbf37238ea5fba19323dacaca982810dbf18387fe139a7555f003e460f9304"} Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.662421 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.824623 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-inventory\") pod \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.824860 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-secret-0\") pod \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.824913 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-combined-ca-bundle\") pod \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.825016 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-ssh-key-openstack-edpm-ipam\") pod \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.825076 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvkbz\" (UniqueName: \"kubernetes.io/projected/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-kube-api-access-cvkbz\") pod \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\" (UID: \"bcbdff51-5cc0-4011-a1ae-a260ee565ef7\") " Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.831311 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-kube-api-access-cvkbz" (OuterVolumeSpecName: "kube-api-access-cvkbz") pod "bcbdff51-5cc0-4011-a1ae-a260ee565ef7" (UID: "bcbdff51-5cc0-4011-a1ae-a260ee565ef7"). InnerVolumeSpecName "kube-api-access-cvkbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.834037 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bcbdff51-5cc0-4011-a1ae-a260ee565ef7" (UID: "bcbdff51-5cc0-4011-a1ae-a260ee565ef7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.859159 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bcbdff51-5cc0-4011-a1ae-a260ee565ef7" (UID: "bcbdff51-5cc0-4011-a1ae-a260ee565ef7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.875237 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "bcbdff51-5cc0-4011-a1ae-a260ee565ef7" (UID: "bcbdff51-5cc0-4011-a1ae-a260ee565ef7"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.884819 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-inventory" (OuterVolumeSpecName: "inventory") pod "bcbdff51-5cc0-4011-a1ae-a260ee565ef7" (UID: "bcbdff51-5cc0-4011-a1ae-a260ee565ef7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.927690 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvkbz\" (UniqueName: \"kubernetes.io/projected/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-kube-api-access-cvkbz\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.927746 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.927766 4750 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.927783 4750 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:16 crc kubenswrapper[4750]: I0309 19:08:16.927801 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcbdff51-5cc0-4011-a1ae-a260ee565ef7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.106030 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" event={"ID":"bcbdff51-5cc0-4011-a1ae-a260ee565ef7","Type":"ContainerDied","Data":"e964870eb9fa9f8152ba6cdd3b7cff69bf2af4f2cd923b15571ee11767bff1f7"} Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.106098 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e964870eb9fa9f8152ba6cdd3b7cff69bf2af4f2cd923b15571ee11767bff1f7" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.106117 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-827lf" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.218749 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm"] Mar 09 19:08:17 crc kubenswrapper[4750]: E0309 19:08:17.219233 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcbdff51-5cc0-4011-a1ae-a260ee565ef7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.219254 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcbdff51-5cc0-4011-a1ae-a260ee565ef7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 09 19:08:17 crc kubenswrapper[4750]: E0309 19:08:17.219311 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="160d1e84-3414-4465-8ad7-d142374304aa" containerName="oc" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.219323 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="160d1e84-3414-4465-8ad7-d142374304aa" containerName="oc" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.219548 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="160d1e84-3414-4465-8ad7-d142374304aa" containerName="oc" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.219584 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcbdff51-5cc0-4011-a1ae-a260ee565ef7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.220482 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.224198 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.224274 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.224477 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.224529 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.224653 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.224536 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.224765 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.245689 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm"] Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.340566 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.340618 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.340745 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.340801 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vdw6\" (UniqueName: \"kubernetes.io/projected/f8f28da2-4627-4099-8838-4ac2b22c2209-kube-api-access-9vdw6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.340913 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.341090 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.341116 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.341176 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.341198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.341244 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.341312 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442734 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442786 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vdw6\" (UniqueName: \"kubernetes.io/projected/f8f28da2-4627-4099-8838-4ac2b22c2209-kube-api-access-9vdw6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442832 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442873 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442894 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442920 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442938 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442961 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.442988 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.443030 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.443051 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.444037 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.446751 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.447598 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.447911 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.447996 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.448007 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.448365 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.448590 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.448787 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.449473 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.461096 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vdw6\" (UniqueName: \"kubernetes.io/projected/f8f28da2-4627-4099-8838-4ac2b22c2209-kube-api-access-9vdw6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xlzzm\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:17 crc kubenswrapper[4750]: I0309 19:08:17.537696 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:08:18 crc kubenswrapper[4750]: I0309 19:08:18.116785 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm"] Mar 09 19:08:18 crc kubenswrapper[4750]: I0309 19:08:18.964078 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-87gz7"] Mar 09 19:08:18 crc kubenswrapper[4750]: I0309 19:08:18.966980 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:18 crc kubenswrapper[4750]: I0309 19:08:18.975370 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-87gz7"] Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.080022 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-utilities\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.080123 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6m69\" (UniqueName: \"kubernetes.io/projected/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-kube-api-access-h6m69\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.080211 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-catalog-content\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.136475 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" event={"ID":"f8f28da2-4627-4099-8838-4ac2b22c2209","Type":"ContainerStarted","Data":"51a989eb1099eab2c845a84c04cc9695665442fc8e95ca1d5e93498849bc926e"} Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.136525 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" event={"ID":"f8f28da2-4627-4099-8838-4ac2b22c2209","Type":"ContainerStarted","Data":"cb775cac58c9e94d2f01b78b80e69dd9ea62b73ef5ec6da82dbf16f472dff730"} Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.158182 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" podStartSLOduration=1.709528395 podStartE2EDuration="2.158164748s" podCreationTimestamp="2026-03-09 19:08:17 +0000 UTC" firstStartedPulling="2026-03-09 19:08:18.120102963 +0000 UTC m=+2579.462575361" lastFinishedPulling="2026-03-09 19:08:18.568739306 +0000 UTC m=+2579.911211714" observedRunningTime="2026-03-09 19:08:19.15091167 +0000 UTC m=+2580.493384078" watchObservedRunningTime="2026-03-09 19:08:19.158164748 +0000 UTC m=+2580.500637156" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.181760 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-catalog-content\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.181832 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-utilities\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.181922 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6m69\" (UniqueName: \"kubernetes.io/projected/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-kube-api-access-h6m69\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.182603 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-catalog-content\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.182703 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-utilities\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.207284 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6m69\" (UniqueName: \"kubernetes.io/projected/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-kube-api-access-h6m69\") pod \"community-operators-87gz7\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.327317 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:19 crc kubenswrapper[4750]: I0309 19:08:19.862267 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-87gz7"] Mar 09 19:08:19 crc kubenswrapper[4750]: W0309 19:08:19.868451 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae957bba_9c1c_43c5_a16a_9d6676e2ba17.slice/crio-a6ab0fa492ad766fee40f95280f025fc7484cf7c0207eb10bd68e27d1db9c0fd WatchSource:0}: Error finding container a6ab0fa492ad766fee40f95280f025fc7484cf7c0207eb10bd68e27d1db9c0fd: Status 404 returned error can't find the container with id a6ab0fa492ad766fee40f95280f025fc7484cf7c0207eb10bd68e27d1db9c0fd Mar 09 19:08:20 crc kubenswrapper[4750]: I0309 19:08:20.147347 4750 generic.go:334] "Generic (PLEG): container finished" podID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerID="616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0" exitCode=0 Mar 09 19:08:20 crc kubenswrapper[4750]: I0309 19:08:20.147515 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87gz7" event={"ID":"ae957bba-9c1c-43c5-a16a-9d6676e2ba17","Type":"ContainerDied","Data":"616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0"} Mar 09 19:08:20 crc kubenswrapper[4750]: I0309 19:08:20.148339 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87gz7" event={"ID":"ae957bba-9c1c-43c5-a16a-9d6676e2ba17","Type":"ContainerStarted","Data":"a6ab0fa492ad766fee40f95280f025fc7484cf7c0207eb10bd68e27d1db9c0fd"} Mar 09 19:08:20 crc kubenswrapper[4750]: I0309 19:08:20.150110 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:08:22 crc kubenswrapper[4750]: I0309 19:08:22.172967 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87gz7" event={"ID":"ae957bba-9c1c-43c5-a16a-9d6676e2ba17","Type":"ContainerStarted","Data":"205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f"} Mar 09 19:08:23 crc kubenswrapper[4750]: I0309 19:08:23.182507 4750 generic.go:334] "Generic (PLEG): container finished" podID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerID="205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f" exitCode=0 Mar 09 19:08:23 crc kubenswrapper[4750]: I0309 19:08:23.182802 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87gz7" event={"ID":"ae957bba-9c1c-43c5-a16a-9d6676e2ba17","Type":"ContainerDied","Data":"205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f"} Mar 09 19:08:24 crc kubenswrapper[4750]: I0309 19:08:24.193915 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87gz7" event={"ID":"ae957bba-9c1c-43c5-a16a-9d6676e2ba17","Type":"ContainerStarted","Data":"9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba"} Mar 09 19:08:24 crc kubenswrapper[4750]: I0309 19:08:24.220246 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-87gz7" podStartSLOduration=2.634701068 podStartE2EDuration="6.220229477s" podCreationTimestamp="2026-03-09 19:08:18 +0000 UTC" firstStartedPulling="2026-03-09 19:08:20.149915047 +0000 UTC m=+2581.492387435" lastFinishedPulling="2026-03-09 19:08:23.735443456 +0000 UTC m=+2585.077915844" observedRunningTime="2026-03-09 19:08:24.212922837 +0000 UTC m=+2585.555395235" watchObservedRunningTime="2026-03-09 19:08:24.220229477 +0000 UTC m=+2585.562701875" Mar 09 19:08:26 crc kubenswrapper[4750]: I0309 19:08:26.374710 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:08:26 crc kubenswrapper[4750]: E0309 19:08:26.375037 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:08:29 crc kubenswrapper[4750]: I0309 19:08:29.328121 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:29 crc kubenswrapper[4750]: I0309 19:08:29.328445 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:29 crc kubenswrapper[4750]: I0309 19:08:29.420617 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:30 crc kubenswrapper[4750]: I0309 19:08:30.331271 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:31 crc kubenswrapper[4750]: I0309 19:08:31.568735 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-87gz7"] Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.275015 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-87gz7" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="registry-server" containerID="cri-o://9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba" gracePeriod=2 Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.798840 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.897697 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-utilities\") pod \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.897809 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-catalog-content\") pod \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.898010 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6m69\" (UniqueName: \"kubernetes.io/projected/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-kube-api-access-h6m69\") pod \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\" (UID: \"ae957bba-9c1c-43c5-a16a-9d6676e2ba17\") " Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.900886 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-utilities" (OuterVolumeSpecName: "utilities") pod "ae957bba-9c1c-43c5-a16a-9d6676e2ba17" (UID: "ae957bba-9c1c-43c5-a16a-9d6676e2ba17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.911334 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-kube-api-access-h6m69" (OuterVolumeSpecName: "kube-api-access-h6m69") pod "ae957bba-9c1c-43c5-a16a-9d6676e2ba17" (UID: "ae957bba-9c1c-43c5-a16a-9d6676e2ba17"). InnerVolumeSpecName "kube-api-access-h6m69". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:08:32 crc kubenswrapper[4750]: I0309 19:08:32.980164 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae957bba-9c1c-43c5-a16a-9d6676e2ba17" (UID: "ae957bba-9c1c-43c5-a16a-9d6676e2ba17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.000783 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6m69\" (UniqueName: \"kubernetes.io/projected/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-kube-api-access-h6m69\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.000829 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.000842 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae957bba-9c1c-43c5-a16a-9d6676e2ba17-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.289947 4750 generic.go:334] "Generic (PLEG): container finished" podID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerID="9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba" exitCode=0 Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.289995 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87gz7" event={"ID":"ae957bba-9c1c-43c5-a16a-9d6676e2ba17","Type":"ContainerDied","Data":"9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba"} Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.290028 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87gz7" event={"ID":"ae957bba-9c1c-43c5-a16a-9d6676e2ba17","Type":"ContainerDied","Data":"a6ab0fa492ad766fee40f95280f025fc7484cf7c0207eb10bd68e27d1db9c0fd"} Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.290041 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-87gz7" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.290050 4750 scope.go:117] "RemoveContainer" containerID="9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.331324 4750 scope.go:117] "RemoveContainer" containerID="205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.349317 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-87gz7"] Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.362897 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-87gz7"] Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.364829 4750 scope.go:117] "RemoveContainer" containerID="616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.388002 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" path="/var/lib/kubelet/pods/ae957bba-9c1c-43c5-a16a-9d6676e2ba17/volumes" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.435314 4750 scope.go:117] "RemoveContainer" containerID="9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba" Mar 09 19:08:33 crc kubenswrapper[4750]: E0309 19:08:33.435975 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba\": container with ID starting with 9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba not found: ID does not exist" containerID="9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.436028 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba"} err="failed to get container status \"9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba\": rpc error: code = NotFound desc = could not find container \"9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba\": container with ID starting with 9e44c00c2f7baef0d6e7171bf8c307d45cf0e9cc882bb4336a591252b3ad64ba not found: ID does not exist" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.436062 4750 scope.go:117] "RemoveContainer" containerID="205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f" Mar 09 19:08:33 crc kubenswrapper[4750]: E0309 19:08:33.436439 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f\": container with ID starting with 205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f not found: ID does not exist" containerID="205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.436483 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f"} err="failed to get container status \"205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f\": rpc error: code = NotFound desc = could not find container \"205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f\": container with ID starting with 205d6cde261289d59e3bbbca072cabf849dca3d955899c70cc4bebe2a34af99f not found: ID does not exist" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.436508 4750 scope.go:117] "RemoveContainer" containerID="616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0" Mar 09 19:08:33 crc kubenswrapper[4750]: E0309 19:08:33.436928 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0\": container with ID starting with 616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0 not found: ID does not exist" containerID="616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0" Mar 09 19:08:33 crc kubenswrapper[4750]: I0309 19:08:33.436970 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0"} err="failed to get container status \"616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0\": rpc error: code = NotFound desc = could not find container \"616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0\": container with ID starting with 616f86d36e5a1beb3275f285352d33dd2dd7c5481dd61d711fd90494c28252f0 not found: ID does not exist" Mar 09 19:08:37 crc kubenswrapper[4750]: I0309 19:08:37.374408 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:08:37 crc kubenswrapper[4750]: E0309 19:08:37.375476 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:08:45 crc kubenswrapper[4750]: I0309 19:08:45.500326 4750 scope.go:117] "RemoveContainer" containerID="5ae8055273441e853eb425756a9530ec7cd8d44b0f8d53ca4e59bb29e5dff8c6" Mar 09 19:08:51 crc kubenswrapper[4750]: I0309 19:08:51.373619 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:08:51 crc kubenswrapper[4750]: E0309 19:08:51.374353 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:09:03 crc kubenswrapper[4750]: I0309 19:09:03.374301 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:09:03 crc kubenswrapper[4750]: E0309 19:09:03.375555 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:09:15 crc kubenswrapper[4750]: I0309 19:09:15.373911 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:09:15 crc kubenswrapper[4750]: E0309 19:09:15.375312 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:09:28 crc kubenswrapper[4750]: I0309 19:09:28.373710 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:09:28 crc kubenswrapper[4750]: E0309 19:09:28.376450 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:09:40 crc kubenswrapper[4750]: I0309 19:09:40.374420 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:09:40 crc kubenswrapper[4750]: E0309 19:09:40.375491 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:09:51 crc kubenswrapper[4750]: I0309 19:09:51.374836 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:09:51 crc kubenswrapper[4750]: E0309 19:09:51.376231 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.171163 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551390-xzz7r"] Mar 09 19:10:00 crc kubenswrapper[4750]: E0309 19:10:00.172559 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="extract-content" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.172580 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="extract-content" Mar 09 19:10:00 crc kubenswrapper[4750]: E0309 19:10:00.172609 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="extract-utilities" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.172622 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="extract-utilities" Mar 09 19:10:00 crc kubenswrapper[4750]: E0309 19:10:00.172683 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="registry-server" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.172699 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="registry-server" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.173061 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae957bba-9c1c-43c5-a16a-9d6676e2ba17" containerName="registry-server" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.174260 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.181657 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.182006 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.182478 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.187474 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551390-xzz7r"] Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.292086 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh6dn\" (UniqueName: \"kubernetes.io/projected/87d71783-700a-40fd-96f1-ad7f861ccf38-kube-api-access-gh6dn\") pod \"auto-csr-approver-29551390-xzz7r\" (UID: \"87d71783-700a-40fd-96f1-ad7f861ccf38\") " pod="openshift-infra/auto-csr-approver-29551390-xzz7r" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.393989 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh6dn\" (UniqueName: \"kubernetes.io/projected/87d71783-700a-40fd-96f1-ad7f861ccf38-kube-api-access-gh6dn\") pod \"auto-csr-approver-29551390-xzz7r\" (UID: \"87d71783-700a-40fd-96f1-ad7f861ccf38\") " pod="openshift-infra/auto-csr-approver-29551390-xzz7r" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.416569 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh6dn\" (UniqueName: \"kubernetes.io/projected/87d71783-700a-40fd-96f1-ad7f861ccf38-kube-api-access-gh6dn\") pod \"auto-csr-approver-29551390-xzz7r\" (UID: \"87d71783-700a-40fd-96f1-ad7f861ccf38\") " pod="openshift-infra/auto-csr-approver-29551390-xzz7r" Mar 09 19:10:00 crc kubenswrapper[4750]: I0309 19:10:00.510710 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" Mar 09 19:10:01 crc kubenswrapper[4750]: I0309 19:10:01.046204 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551390-xzz7r"] Mar 09 19:10:02 crc kubenswrapper[4750]: I0309 19:10:02.037035 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" event={"ID":"87d71783-700a-40fd-96f1-ad7f861ccf38","Type":"ContainerStarted","Data":"2ab48a8995b1e58c9fe20b69b12b1bf91621a4843c781c38cefff6cc2626a71d"} Mar 09 19:10:03 crc kubenswrapper[4750]: I0309 19:10:03.045876 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" event={"ID":"87d71783-700a-40fd-96f1-ad7f861ccf38","Type":"ContainerStarted","Data":"07ea2e62d5b33a075bbc264fe9ae81afb8e75d7c1e339f7f970ac954b33e7f4c"} Mar 09 19:10:03 crc kubenswrapper[4750]: I0309 19:10:03.064978 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" podStartSLOduration=1.553288676 podStartE2EDuration="3.064961946s" podCreationTimestamp="2026-03-09 19:10:00 +0000 UTC" firstStartedPulling="2026-03-09 19:10:01.051089979 +0000 UTC m=+2682.393562377" lastFinishedPulling="2026-03-09 19:10:02.562763219 +0000 UTC m=+2683.905235647" observedRunningTime="2026-03-09 19:10:03.05806893 +0000 UTC m=+2684.400541348" watchObservedRunningTime="2026-03-09 19:10:03.064961946 +0000 UTC m=+2684.407434344" Mar 09 19:10:04 crc kubenswrapper[4750]: I0309 19:10:04.059044 4750 generic.go:334] "Generic (PLEG): container finished" podID="87d71783-700a-40fd-96f1-ad7f861ccf38" containerID="07ea2e62d5b33a075bbc264fe9ae81afb8e75d7c1e339f7f970ac954b33e7f4c" exitCode=0 Mar 09 19:10:04 crc kubenswrapper[4750]: I0309 19:10:04.059112 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" event={"ID":"87d71783-700a-40fd-96f1-ad7f861ccf38","Type":"ContainerDied","Data":"07ea2e62d5b33a075bbc264fe9ae81afb8e75d7c1e339f7f970ac954b33e7f4c"} Mar 09 19:10:05 crc kubenswrapper[4750]: I0309 19:10:05.527399 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" Mar 09 19:10:05 crc kubenswrapper[4750]: I0309 19:10:05.618889 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh6dn\" (UniqueName: \"kubernetes.io/projected/87d71783-700a-40fd-96f1-ad7f861ccf38-kube-api-access-gh6dn\") pod \"87d71783-700a-40fd-96f1-ad7f861ccf38\" (UID: \"87d71783-700a-40fd-96f1-ad7f861ccf38\") " Mar 09 19:10:05 crc kubenswrapper[4750]: I0309 19:10:05.625814 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87d71783-700a-40fd-96f1-ad7f861ccf38-kube-api-access-gh6dn" (OuterVolumeSpecName: "kube-api-access-gh6dn") pod "87d71783-700a-40fd-96f1-ad7f861ccf38" (UID: "87d71783-700a-40fd-96f1-ad7f861ccf38"). InnerVolumeSpecName "kube-api-access-gh6dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:10:05 crc kubenswrapper[4750]: I0309 19:10:05.721666 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh6dn\" (UniqueName: \"kubernetes.io/projected/87d71783-700a-40fd-96f1-ad7f861ccf38-kube-api-access-gh6dn\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:06 crc kubenswrapper[4750]: I0309 19:10:06.085668 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" event={"ID":"87d71783-700a-40fd-96f1-ad7f861ccf38","Type":"ContainerDied","Data":"2ab48a8995b1e58c9fe20b69b12b1bf91621a4843c781c38cefff6cc2626a71d"} Mar 09 19:10:06 crc kubenswrapper[4750]: I0309 19:10:06.086138 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ab48a8995b1e58c9fe20b69b12b1bf91621a4843c781c38cefff6cc2626a71d" Mar 09 19:10:06 crc kubenswrapper[4750]: I0309 19:10:06.085766 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551390-xzz7r" Mar 09 19:10:06 crc kubenswrapper[4750]: I0309 19:10:06.147042 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551384-n8pxf"] Mar 09 19:10:06 crc kubenswrapper[4750]: I0309 19:10:06.157179 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551384-n8pxf"] Mar 09 19:10:06 crc kubenswrapper[4750]: I0309 19:10:06.375138 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:10:06 crc kubenswrapper[4750]: E0309 19:10:06.375670 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:10:07 crc kubenswrapper[4750]: I0309 19:10:07.392924 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="010e3beb-8423-4ab0-b4b4-5f3344e594ed" path="/var/lib/kubelet/pods/010e3beb-8423-4ab0-b4b4-5f3344e594ed/volumes" Mar 09 19:10:20 crc kubenswrapper[4750]: I0309 19:10:20.373790 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:10:20 crc kubenswrapper[4750]: E0309 19:10:20.374778 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.472896 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8bn66"] Mar 09 19:10:25 crc kubenswrapper[4750]: E0309 19:10:25.474832 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87d71783-700a-40fd-96f1-ad7f861ccf38" containerName="oc" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.474932 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="87d71783-700a-40fd-96f1-ad7f861ccf38" containerName="oc" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.475266 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="87d71783-700a-40fd-96f1-ad7f861ccf38" containerName="oc" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.479908 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.501568 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8bn66"] Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.581828 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwkvf\" (UniqueName: \"kubernetes.io/projected/2830f628-c59c-4704-ad18-ae6aeab3d83a-kube-api-access-qwkvf\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.581889 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-catalog-content\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.581920 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-utilities\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.683680 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwkvf\" (UniqueName: \"kubernetes.io/projected/2830f628-c59c-4704-ad18-ae6aeab3d83a-kube-api-access-qwkvf\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.683735 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-catalog-content\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.683763 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-utilities\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.684325 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-utilities\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.684409 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-catalog-content\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.712947 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwkvf\" (UniqueName: \"kubernetes.io/projected/2830f628-c59c-4704-ad18-ae6aeab3d83a-kube-api-access-qwkvf\") pod \"redhat-operators-8bn66\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:25 crc kubenswrapper[4750]: I0309 19:10:25.810755 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:26 crc kubenswrapper[4750]: I0309 19:10:26.286826 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8bn66"] Mar 09 19:10:26 crc kubenswrapper[4750]: I0309 19:10:26.327501 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bn66" event={"ID":"2830f628-c59c-4704-ad18-ae6aeab3d83a","Type":"ContainerStarted","Data":"71dc039639c4b50f3216a0783b7a44bb164741da89df29658e39b279a56262b4"} Mar 09 19:10:27 crc kubenswrapper[4750]: I0309 19:10:27.338299 4750 generic.go:334] "Generic (PLEG): container finished" podID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerID="4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099" exitCode=0 Mar 09 19:10:27 crc kubenswrapper[4750]: I0309 19:10:27.338506 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bn66" event={"ID":"2830f628-c59c-4704-ad18-ae6aeab3d83a","Type":"ContainerDied","Data":"4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099"} Mar 09 19:10:28 crc kubenswrapper[4750]: I0309 19:10:28.356226 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bn66" event={"ID":"2830f628-c59c-4704-ad18-ae6aeab3d83a","Type":"ContainerStarted","Data":"328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c"} Mar 09 19:10:32 crc kubenswrapper[4750]: I0309 19:10:32.374099 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:10:32 crc kubenswrapper[4750]: E0309 19:10:32.375112 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:10:33 crc kubenswrapper[4750]: I0309 19:10:33.415118 4750 generic.go:334] "Generic (PLEG): container finished" podID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerID="328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c" exitCode=0 Mar 09 19:10:33 crc kubenswrapper[4750]: I0309 19:10:33.415451 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bn66" event={"ID":"2830f628-c59c-4704-ad18-ae6aeab3d83a","Type":"ContainerDied","Data":"328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c"} Mar 09 19:10:35 crc kubenswrapper[4750]: I0309 19:10:35.437682 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bn66" event={"ID":"2830f628-c59c-4704-ad18-ae6aeab3d83a","Type":"ContainerStarted","Data":"bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5"} Mar 09 19:10:35 crc kubenswrapper[4750]: I0309 19:10:35.458598 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8bn66" podStartSLOduration=2.739017141 podStartE2EDuration="10.458578606s" podCreationTimestamp="2026-03-09 19:10:25 +0000 UTC" firstStartedPulling="2026-03-09 19:10:27.341305571 +0000 UTC m=+2708.683777969" lastFinishedPulling="2026-03-09 19:10:35.060867006 +0000 UTC m=+2716.403339434" observedRunningTime="2026-03-09 19:10:35.45393224 +0000 UTC m=+2716.796404668" watchObservedRunningTime="2026-03-09 19:10:35.458578606 +0000 UTC m=+2716.801051014" Mar 09 19:10:35 crc kubenswrapper[4750]: I0309 19:10:35.811024 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:35 crc kubenswrapper[4750]: I0309 19:10:35.811090 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:36 crc kubenswrapper[4750]: I0309 19:10:36.882183 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8bn66" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="registry-server" probeResult="failure" output=< Mar 09 19:10:36 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:10:36 crc kubenswrapper[4750]: > Mar 09 19:10:45 crc kubenswrapper[4750]: I0309 19:10:45.656022 4750 scope.go:117] "RemoveContainer" containerID="18852841a2a023c9eb2f2433a02cfe1c2cca08ca9365dd8b34296418cafa21f0" Mar 09 19:10:46 crc kubenswrapper[4750]: I0309 19:10:46.872178 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8bn66" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="registry-server" probeResult="failure" output=< Mar 09 19:10:46 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:10:46 crc kubenswrapper[4750]: > Mar 09 19:10:47 crc kubenswrapper[4750]: I0309 19:10:47.374383 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:10:47 crc kubenswrapper[4750]: E0309 19:10:47.374694 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:10:54 crc kubenswrapper[4750]: I0309 19:10:54.670586 4750 generic.go:334] "Generic (PLEG): container finished" podID="f8f28da2-4627-4099-8838-4ac2b22c2209" containerID="51a989eb1099eab2c845a84c04cc9695665442fc8e95ca1d5e93498849bc926e" exitCode=0 Mar 09 19:10:54 crc kubenswrapper[4750]: I0309 19:10:54.670721 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" event={"ID":"f8f28da2-4627-4099-8838-4ac2b22c2209","Type":"ContainerDied","Data":"51a989eb1099eab2c845a84c04cc9695665442fc8e95ca1d5e93498849bc926e"} Mar 09 19:10:55 crc kubenswrapper[4750]: I0309 19:10:55.929049 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:55 crc kubenswrapper[4750]: I0309 19:10:55.991890 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.108395 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.224943 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-extra-config-0\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.224993 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-0\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225041 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-2\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225131 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-1\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225189 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-3\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225229 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-combined-ca-bundle\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225280 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vdw6\" (UniqueName: \"kubernetes.io/projected/f8f28da2-4627-4099-8838-4ac2b22c2209-kube-api-access-9vdw6\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225351 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-inventory\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225459 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-0\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225487 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-1\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.225523 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-ssh-key-openstack-edpm-ipam\") pod \"f8f28da2-4627-4099-8838-4ac2b22c2209\" (UID: \"f8f28da2-4627-4099-8838-4ac2b22c2209\") " Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.231516 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f28da2-4627-4099-8838-4ac2b22c2209-kube-api-access-9vdw6" (OuterVolumeSpecName: "kube-api-access-9vdw6") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "kube-api-access-9vdw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.232828 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.253621 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.255940 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.267516 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-inventory" (OuterVolumeSpecName: "inventory") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.268166 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.272501 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.273081 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.275852 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.276321 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.291552 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "f8f28da2-4627-4099-8838-4ac2b22c2209" (UID: "f8f28da2-4627-4099-8838-4ac2b22c2209"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.327975 4750 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328008 4750 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328055 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328159 4750 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328173 4750 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328181 4750 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328190 4750 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328200 4750 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328208 4750 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328216 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vdw6\" (UniqueName: \"kubernetes.io/projected/f8f28da2-4627-4099-8838-4ac2b22c2209-kube-api-access-9vdw6\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.328234 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8f28da2-4627-4099-8838-4ac2b22c2209-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.678757 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8bn66"] Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.694920 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" event={"ID":"f8f28da2-4627-4099-8838-4ac2b22c2209","Type":"ContainerDied","Data":"cb775cac58c9e94d2f01b78b80e69dd9ea62b73ef5ec6da82dbf16f472dff730"} Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.694991 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb775cac58c9e94d2f01b78b80e69dd9ea62b73ef5ec6da82dbf16f472dff730" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.695162 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xlzzm" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.820750 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd"] Mar 09 19:10:56 crc kubenswrapper[4750]: E0309 19:10:56.821458 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f28da2-4627-4099-8838-4ac2b22c2209" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.821545 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f28da2-4627-4099-8838-4ac2b22c2209" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.822231 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f28da2-4627-4099-8838-4ac2b22c2209" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.823462 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.825255 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.826509 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tqmmk" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.827315 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.827797 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.828124 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.844561 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd"] Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.941427 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.941518 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.941547 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.941564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8wqx\" (UniqueName: \"kubernetes.io/projected/c0cdad8f-0238-4da8-94ae-39b139b83613-kube-api-access-w8wqx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.941598 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.941645 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:56 crc kubenswrapper[4750]: I0309 19:10:56.941668 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.044117 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.044184 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8wqx\" (UniqueName: \"kubernetes.io/projected/c0cdad8f-0238-4da8-94ae-39b139b83613-kube-api-access-w8wqx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.044244 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.044287 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.044328 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.044514 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.044661 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.052450 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.052450 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.053908 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.054422 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.062894 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.073608 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.074453 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8wqx\" (UniqueName: \"kubernetes.io/projected/c0cdad8f-0238-4da8-94ae-39b139b83613-kube-api-access-w8wqx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2mknd\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.147729 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.703725 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8bn66" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="registry-server" containerID="cri-o://bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5" gracePeriod=2 Mar 09 19:10:57 crc kubenswrapper[4750]: I0309 19:10:57.715338 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd"] Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.375846 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.475573 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-catalog-content\") pod \"2830f628-c59c-4704-ad18-ae6aeab3d83a\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.475881 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwkvf\" (UniqueName: \"kubernetes.io/projected/2830f628-c59c-4704-ad18-ae6aeab3d83a-kube-api-access-qwkvf\") pod \"2830f628-c59c-4704-ad18-ae6aeab3d83a\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.476229 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-utilities\") pod \"2830f628-c59c-4704-ad18-ae6aeab3d83a\" (UID: \"2830f628-c59c-4704-ad18-ae6aeab3d83a\") " Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.478025 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-utilities" (OuterVolumeSpecName: "utilities") pod "2830f628-c59c-4704-ad18-ae6aeab3d83a" (UID: "2830f628-c59c-4704-ad18-ae6aeab3d83a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.482989 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2830f628-c59c-4704-ad18-ae6aeab3d83a-kube-api-access-qwkvf" (OuterVolumeSpecName: "kube-api-access-qwkvf") pod "2830f628-c59c-4704-ad18-ae6aeab3d83a" (UID: "2830f628-c59c-4704-ad18-ae6aeab3d83a"). InnerVolumeSpecName "kube-api-access-qwkvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.578498 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwkvf\" (UniqueName: \"kubernetes.io/projected/2830f628-c59c-4704-ad18-ae6aeab3d83a-kube-api-access-qwkvf\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.578534 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.599639 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2830f628-c59c-4704-ad18-ae6aeab3d83a" (UID: "2830f628-c59c-4704-ad18-ae6aeab3d83a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.680302 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2830f628-c59c-4704-ad18-ae6aeab3d83a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.719408 4750 generic.go:334] "Generic (PLEG): container finished" podID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerID="bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5" exitCode=0 Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.719484 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bn66" event={"ID":"2830f628-c59c-4704-ad18-ae6aeab3d83a","Type":"ContainerDied","Data":"bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5"} Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.719517 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bn66" event={"ID":"2830f628-c59c-4704-ad18-ae6aeab3d83a","Type":"ContainerDied","Data":"71dc039639c4b50f3216a0783b7a44bb164741da89df29658e39b279a56262b4"} Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.719539 4750 scope.go:117] "RemoveContainer" containerID="bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.719735 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bn66" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.722876 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" event={"ID":"c0cdad8f-0238-4da8-94ae-39b139b83613","Type":"ContainerStarted","Data":"605035f5c6d50e03387d789712896300b0c54a0926d721b2b1c94807f3f6a9be"} Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.722910 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" event={"ID":"c0cdad8f-0238-4da8-94ae-39b139b83613","Type":"ContainerStarted","Data":"e611d5e0faf69b7de113cc6791a50bce2fc9bf5f1c134b8ff284ccace1b94ac5"} Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.754029 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" podStartSLOduration=2.24591929 podStartE2EDuration="2.753998416s" podCreationTimestamp="2026-03-09 19:10:56 +0000 UTC" firstStartedPulling="2026-03-09 19:10:57.72500191 +0000 UTC m=+2739.067474318" lastFinishedPulling="2026-03-09 19:10:58.233081046 +0000 UTC m=+2739.575553444" observedRunningTime="2026-03-09 19:10:58.749227039 +0000 UTC m=+2740.091699437" watchObservedRunningTime="2026-03-09 19:10:58.753998416 +0000 UTC m=+2740.096470804" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.765164 4750 scope.go:117] "RemoveContainer" containerID="328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.798816 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8bn66"] Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.815217 4750 scope.go:117] "RemoveContainer" containerID="4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.819435 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8bn66"] Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.838386 4750 scope.go:117] "RemoveContainer" containerID="bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5" Mar 09 19:10:58 crc kubenswrapper[4750]: E0309 19:10:58.839124 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5\": container with ID starting with bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5 not found: ID does not exist" containerID="bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.839177 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5"} err="failed to get container status \"bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5\": rpc error: code = NotFound desc = could not find container \"bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5\": container with ID starting with bac2809de2032fd662e72846208e39d629253d947113c4b01f1ed0148554a4f5 not found: ID does not exist" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.839206 4750 scope.go:117] "RemoveContainer" containerID="328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c" Mar 09 19:10:58 crc kubenswrapper[4750]: E0309 19:10:58.839524 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c\": container with ID starting with 328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c not found: ID does not exist" containerID="328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.839557 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c"} err="failed to get container status \"328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c\": rpc error: code = NotFound desc = could not find container \"328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c\": container with ID starting with 328f10faf8c109f49a75c768eb68872b9cda701cb24240465af748d68572558c not found: ID does not exist" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.839582 4750 scope.go:117] "RemoveContainer" containerID="4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099" Mar 09 19:10:58 crc kubenswrapper[4750]: E0309 19:10:58.839848 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099\": container with ID starting with 4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099 not found: ID does not exist" containerID="4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099" Mar 09 19:10:58 crc kubenswrapper[4750]: I0309 19:10:58.839873 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099"} err="failed to get container status \"4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099\": rpc error: code = NotFound desc = could not find container \"4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099\": container with ID starting with 4005050de5a40dcf3f2aeb153e83d05f6d24815308861adb2748d48b797ea099 not found: ID does not exist" Mar 09 19:10:59 crc kubenswrapper[4750]: I0309 19:10:59.386084 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" path="/var/lib/kubelet/pods/2830f628-c59c-4704-ad18-ae6aeab3d83a/volumes" Mar 09 19:11:02 crc kubenswrapper[4750]: I0309 19:11:02.372980 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:11:02 crc kubenswrapper[4750]: E0309 19:11:02.373547 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:11:16 crc kubenswrapper[4750]: I0309 19:11:16.374559 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:11:16 crc kubenswrapper[4750]: E0309 19:11:16.375770 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:11:28 crc kubenswrapper[4750]: I0309 19:11:28.374158 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:11:29 crc kubenswrapper[4750]: I0309 19:11:29.094602 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"f78f5d76d0e8b5025f0101c0000bfefe714e5ba77c6b39281d9459bc8318a103"} Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.147578 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551392-dqmh4"] Mar 09 19:12:00 crc kubenswrapper[4750]: E0309 19:12:00.148614 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="extract-content" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.148652 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="extract-content" Mar 09 19:12:00 crc kubenswrapper[4750]: E0309 19:12:00.148708 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="registry-server" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.148717 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="registry-server" Mar 09 19:12:00 crc kubenswrapper[4750]: E0309 19:12:00.148752 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="extract-utilities" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.148771 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="extract-utilities" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.149329 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2830f628-c59c-4704-ad18-ae6aeab3d83a" containerName="registry-server" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.152619 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551392-dqmh4" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.155074 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.155173 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.155106 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.158606 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551392-dqmh4"] Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.283756 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvgzw\" (UniqueName: \"kubernetes.io/projected/b1527157-9fda-4756-b0aa-8fa60c57b85f-kube-api-access-xvgzw\") pod \"auto-csr-approver-29551392-dqmh4\" (UID: \"b1527157-9fda-4756-b0aa-8fa60c57b85f\") " pod="openshift-infra/auto-csr-approver-29551392-dqmh4" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.385149 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvgzw\" (UniqueName: \"kubernetes.io/projected/b1527157-9fda-4756-b0aa-8fa60c57b85f-kube-api-access-xvgzw\") pod \"auto-csr-approver-29551392-dqmh4\" (UID: \"b1527157-9fda-4756-b0aa-8fa60c57b85f\") " pod="openshift-infra/auto-csr-approver-29551392-dqmh4" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.403478 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvgzw\" (UniqueName: \"kubernetes.io/projected/b1527157-9fda-4756-b0aa-8fa60c57b85f-kube-api-access-xvgzw\") pod \"auto-csr-approver-29551392-dqmh4\" (UID: \"b1527157-9fda-4756-b0aa-8fa60c57b85f\") " pod="openshift-infra/auto-csr-approver-29551392-dqmh4" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.485066 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551392-dqmh4" Mar 09 19:12:00 crc kubenswrapper[4750]: I0309 19:12:00.997467 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551392-dqmh4"] Mar 09 19:12:01 crc kubenswrapper[4750]: I0309 19:12:01.416247 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551392-dqmh4" event={"ID":"b1527157-9fda-4756-b0aa-8fa60c57b85f","Type":"ContainerStarted","Data":"0267beb6fcbbcfe2a4b4c78728cc56e8ec9d6d66d809af257bcc94f1cdbcafc0"} Mar 09 19:12:03 crc kubenswrapper[4750]: I0309 19:12:03.442175 4750 generic.go:334] "Generic (PLEG): container finished" podID="b1527157-9fda-4756-b0aa-8fa60c57b85f" containerID="e688510fa33c4594f5b624031c2e09eaf5ca3723129ce99e855218e66f924a77" exitCode=0 Mar 09 19:12:03 crc kubenswrapper[4750]: I0309 19:12:03.442242 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551392-dqmh4" event={"ID":"b1527157-9fda-4756-b0aa-8fa60c57b85f","Type":"ContainerDied","Data":"e688510fa33c4594f5b624031c2e09eaf5ca3723129ce99e855218e66f924a77"} Mar 09 19:12:04 crc kubenswrapper[4750]: I0309 19:12:04.844127 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551392-dqmh4" Mar 09 19:12:04 crc kubenswrapper[4750]: I0309 19:12:04.978423 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvgzw\" (UniqueName: \"kubernetes.io/projected/b1527157-9fda-4756-b0aa-8fa60c57b85f-kube-api-access-xvgzw\") pod \"b1527157-9fda-4756-b0aa-8fa60c57b85f\" (UID: \"b1527157-9fda-4756-b0aa-8fa60c57b85f\") " Mar 09 19:12:04 crc kubenswrapper[4750]: I0309 19:12:04.985686 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1527157-9fda-4756-b0aa-8fa60c57b85f-kube-api-access-xvgzw" (OuterVolumeSpecName: "kube-api-access-xvgzw") pod "b1527157-9fda-4756-b0aa-8fa60c57b85f" (UID: "b1527157-9fda-4756-b0aa-8fa60c57b85f"). InnerVolumeSpecName "kube-api-access-xvgzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:12:05 crc kubenswrapper[4750]: I0309 19:12:05.082129 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvgzw\" (UniqueName: \"kubernetes.io/projected/b1527157-9fda-4756-b0aa-8fa60c57b85f-kube-api-access-xvgzw\") on node \"crc\" DevicePath \"\"" Mar 09 19:12:05 crc kubenswrapper[4750]: I0309 19:12:05.465191 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551392-dqmh4" event={"ID":"b1527157-9fda-4756-b0aa-8fa60c57b85f","Type":"ContainerDied","Data":"0267beb6fcbbcfe2a4b4c78728cc56e8ec9d6d66d809af257bcc94f1cdbcafc0"} Mar 09 19:12:05 crc kubenswrapper[4750]: I0309 19:12:05.465534 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0267beb6fcbbcfe2a4b4c78728cc56e8ec9d6d66d809af257bcc94f1cdbcafc0" Mar 09 19:12:05 crc kubenswrapper[4750]: I0309 19:12:05.465283 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551392-dqmh4" Mar 09 19:12:05 crc kubenswrapper[4750]: I0309 19:12:05.950269 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551386-ssglf"] Mar 09 19:12:05 crc kubenswrapper[4750]: I0309 19:12:05.961702 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551386-ssglf"] Mar 09 19:12:07 crc kubenswrapper[4750]: I0309 19:12:07.389939 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8c96fa-17c8-48fd-9d78-8155c97a3748" path="/var/lib/kubelet/pods/3b8c96fa-17c8-48fd-9d78-8155c97a3748/volumes" Mar 09 19:12:45 crc kubenswrapper[4750]: I0309 19:12:45.780937 4750 scope.go:117] "RemoveContainer" containerID="6b7a52fe3fe6a8a8fa77132f544b14e8c8230aae776cc01eb98d3d7e20a546e7" Mar 09 19:13:06 crc kubenswrapper[4750]: I0309 19:13:06.147578 4750 generic.go:334] "Generic (PLEG): container finished" podID="c0cdad8f-0238-4da8-94ae-39b139b83613" containerID="605035f5c6d50e03387d789712896300b0c54a0926d721b2b1c94807f3f6a9be" exitCode=0 Mar 09 19:13:06 crc kubenswrapper[4750]: I0309 19:13:06.147721 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" event={"ID":"c0cdad8f-0238-4da8-94ae-39b139b83613","Type":"ContainerDied","Data":"605035f5c6d50e03387d789712896300b0c54a0926d721b2b1c94807f3f6a9be"} Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.693029 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.804647 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-0\") pod \"c0cdad8f-0238-4da8-94ae-39b139b83613\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.804779 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ssh-key-openstack-edpm-ipam\") pod \"c0cdad8f-0238-4da8-94ae-39b139b83613\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.804826 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8wqx\" (UniqueName: \"kubernetes.io/projected/c0cdad8f-0238-4da8-94ae-39b139b83613-kube-api-access-w8wqx\") pod \"c0cdad8f-0238-4da8-94ae-39b139b83613\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.804916 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-2\") pod \"c0cdad8f-0238-4da8-94ae-39b139b83613\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.805009 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-telemetry-combined-ca-bundle\") pod \"c0cdad8f-0238-4da8-94ae-39b139b83613\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.805077 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-inventory\") pod \"c0cdad8f-0238-4da8-94ae-39b139b83613\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.805118 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-1\") pod \"c0cdad8f-0238-4da8-94ae-39b139b83613\" (UID: \"c0cdad8f-0238-4da8-94ae-39b139b83613\") " Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.811553 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c0cdad8f-0238-4da8-94ae-39b139b83613" (UID: "c0cdad8f-0238-4da8-94ae-39b139b83613"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.813995 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0cdad8f-0238-4da8-94ae-39b139b83613-kube-api-access-w8wqx" (OuterVolumeSpecName: "kube-api-access-w8wqx") pod "c0cdad8f-0238-4da8-94ae-39b139b83613" (UID: "c0cdad8f-0238-4da8-94ae-39b139b83613"). InnerVolumeSpecName "kube-api-access-w8wqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.844953 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "c0cdad8f-0238-4da8-94ae-39b139b83613" (UID: "c0cdad8f-0238-4da8-94ae-39b139b83613"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.853604 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-inventory" (OuterVolumeSpecName: "inventory") pod "c0cdad8f-0238-4da8-94ae-39b139b83613" (UID: "c0cdad8f-0238-4da8-94ae-39b139b83613"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.858844 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "c0cdad8f-0238-4da8-94ae-39b139b83613" (UID: "c0cdad8f-0238-4da8-94ae-39b139b83613"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.866897 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "c0cdad8f-0238-4da8-94ae-39b139b83613" (UID: "c0cdad8f-0238-4da8-94ae-39b139b83613"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.874272 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c0cdad8f-0238-4da8-94ae-39b139b83613" (UID: "c0cdad8f-0238-4da8-94ae-39b139b83613"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.908746 4750 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.908816 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8wqx\" (UniqueName: \"kubernetes.io/projected/c0cdad8f-0238-4da8-94ae-39b139b83613-kube-api-access-w8wqx\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.908838 4750 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.908859 4750 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.908879 4750 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-inventory\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.908896 4750 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:07 crc kubenswrapper[4750]: I0309 19:13:07.908914 4750 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c0cdad8f-0238-4da8-94ae-39b139b83613-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:08 crc kubenswrapper[4750]: I0309 19:13:08.171707 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" event={"ID":"c0cdad8f-0238-4da8-94ae-39b139b83613","Type":"ContainerDied","Data":"e611d5e0faf69b7de113cc6791a50bce2fc9bf5f1c134b8ff284ccace1b94ac5"} Mar 09 19:13:08 crc kubenswrapper[4750]: I0309 19:13:08.172026 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e611d5e0faf69b7de113cc6791a50bce2fc9bf5f1c134b8ff284ccace1b94ac5" Mar 09 19:13:08 crc kubenswrapper[4750]: I0309 19:13:08.172081 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2mknd" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.307154 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.307784 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="cinder-scheduler" containerID="cri-o://76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2" gracePeriod=30 Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.308207 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="probe" containerID="cri-o://65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9" gracePeriod=30 Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.412902 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Mar 09 19:13:42 crc kubenswrapper[4750]: E0309 19:13:42.413364 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0cdad8f-0238-4da8-94ae-39b139b83613" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.413391 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0cdad8f-0238-4da8-94ae-39b139b83613" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 09 19:13:42 crc kubenswrapper[4750]: E0309 19:13:42.413418 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1527157-9fda-4756-b0aa-8fa60c57b85f" containerName="oc" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.413426 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1527157-9fda-4756-b0aa-8fa60c57b85f" containerName="oc" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.413695 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1527157-9fda-4756-b0aa-8fa60c57b85f" containerName="oc" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.413729 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0cdad8f-0238-4da8-94ae-39b139b83613" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.418872 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.421169 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.429157 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.532081 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-0"] Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.533627 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.536748 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-config-data" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.564930 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.587376 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.589160 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.597122 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-2-config-data" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605016 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-sys\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605060 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-dev\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605087 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605124 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-scripts\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605143 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605180 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-lib-modules\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605219 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605237 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69vv6\" (UniqueName: \"kubernetes.io/projected/dbe58b04-c73c-4292-825f-1e6b466ad54f-kube-api-access-69vv6\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605266 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605288 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605304 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-run\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605337 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605354 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.605376 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-config-data\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.615265 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.706852 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.706917 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.706939 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.706965 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-sys\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.706985 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707096 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-sys\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707158 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707208 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-dev\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707252 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707263 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-dev\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707283 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707342 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwwnn\" (UniqueName: \"kubernetes.io/projected/e27746ea-410c-45a2-91f5-58633d7c7747-kube-api-access-zwwnn\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707369 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707403 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-scripts\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707438 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707454 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707481 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707502 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707741 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5pmq\" (UniqueName: \"kubernetes.io/projected/a22fae28-039f-4b62-b889-f53da0ef827b-kube-api-access-w5pmq\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707835 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-dev\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707903 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.707979 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-run\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708014 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708081 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-lib-modules\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708143 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708205 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708235 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708198 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708274 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-lib-modules\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708327 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708366 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69vv6\" (UniqueName: \"kubernetes.io/projected/dbe58b04-c73c-4292-825f-1e6b466ad54f-kube-api-access-69vv6\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708460 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708524 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708596 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708681 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708726 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708763 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-run\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708806 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708843 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708881 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708939 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708979 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.708991 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709045 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-run\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709079 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709205 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709244 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709280 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-config-data\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709344 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709375 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709403 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-sys\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709431 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709450 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.709671 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dbe58b04-c73c-4292-825f-1e6b466ad54f-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.716301 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-scripts\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.716519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.716945 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-config-data\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.717546 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe58b04-c73c-4292-825f-1e6b466ad54f-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.726745 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69vv6\" (UniqueName: \"kubernetes.io/projected/dbe58b04-c73c-4292-825f-1e6b466ad54f-kube-api-access-69vv6\") pod \"cinder-backup-0\" (UID: \"dbe58b04-c73c-4292-825f-1e6b466ad54f\") " pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.744539 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811195 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811704 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811737 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811761 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811786 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811813 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811840 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811847 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwwnn\" (UniqueName: \"kubernetes.io/projected/e27746ea-410c-45a2-91f5-58633d7c7747-kube-api-access-zwwnn\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812039 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812085 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812104 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812124 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812176 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5pmq\" (UniqueName: \"kubernetes.io/projected/a22fae28-039f-4b62-b889-f53da0ef827b-kube-api-access-w5pmq\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812213 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-dev\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812222 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812242 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811921 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812278 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-run\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812306 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812317 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812342 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811933 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812379 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812394 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812408 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812432 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-run\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812463 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812502 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812517 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812542 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812567 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812594 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812601 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-dev\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812678 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812747 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812794 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-sys\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812820 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.812836 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.814869 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.814930 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.814956 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.814976 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.815053 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-sys\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.815089 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.815116 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811383 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.811949 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a22fae28-039f-4b62-b889-f53da0ef827b-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.816048 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.816205 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e27746ea-410c-45a2-91f5-58633d7c7747-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.819076 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.820808 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.820901 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.821567 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.825967 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.828007 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e27746ea-410c-45a2-91f5-58633d7c7747-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.829165 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.833784 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22fae28-039f-4b62-b889-f53da0ef827b-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.835733 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5pmq\" (UniqueName: \"kubernetes.io/projected/a22fae28-039f-4b62-b889-f53da0ef827b-kube-api-access-w5pmq\") pod \"cinder-volume-nfs-2-0\" (UID: \"a22fae28-039f-4b62-b889-f53da0ef827b\") " pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.838662 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwwnn\" (UniqueName: \"kubernetes.io/projected/e27746ea-410c-45a2-91f5-58633d7c7747-kube-api-access-zwwnn\") pod \"cinder-volume-nfs-0\" (UID: \"e27746ea-410c-45a2-91f5-58633d7c7747\") " pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.868131 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:42 crc kubenswrapper[4750]: I0309 19:13:42.912711 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.324230 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.328005 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:13:43 crc kubenswrapper[4750]: E0309 19:13:43.334204 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadc68d49_e0cb_4adf_abce_56a9eae4f937.slice/crio-conmon-65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9.scope\": RecentStats: unable to find data in memory cache]" Mar 09 19:13:43 crc kubenswrapper[4750]: W0309 19:13:43.485995 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda22fae28_039f_4b62_b889_f53da0ef827b.slice/crio-fe14ff9159556a2fd4621a913dcf54dfe7677a863bdba43fc76b1b60c4126013 WatchSource:0}: Error finding container fe14ff9159556a2fd4621a913dcf54dfe7677a863bdba43fc76b1b60c4126013: Status 404 returned error can't find the container with id fe14ff9159556a2fd4621a913dcf54dfe7677a863bdba43fc76b1b60c4126013 Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.494213 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.596001 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.641501 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"dbe58b04-c73c-4292-825f-1e6b466ad54f","Type":"ContainerStarted","Data":"3047f7450aaabc8d5707fc3f196041e848930b736870034a8a9cafaf4c5617cc"} Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.643510 4750 generic.go:334] "Generic (PLEG): container finished" podID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerID="65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9" exitCode=0 Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.643568 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"adc68d49-e0cb-4adf-abce-56a9eae4f937","Type":"ContainerDied","Data":"65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9"} Mar 09 19:13:43 crc kubenswrapper[4750]: I0309 19:13:43.644544 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"a22fae28-039f-4b62-b889-f53da0ef827b","Type":"ContainerStarted","Data":"fe14ff9159556a2fd4621a913dcf54dfe7677a863bdba43fc76b1b60c4126013"} Mar 09 19:13:43 crc kubenswrapper[4750]: W0309 19:13:43.657843 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode27746ea_410c_45a2_91f5_58633d7c7747.slice/crio-a4c26156a2ca99f389a40be72f500d6e910a2434e270a77532da488bbeba904f WatchSource:0}: Error finding container a4c26156a2ca99f389a40be72f500d6e910a2434e270a77532da488bbeba904f: Status 404 returned error can't find the container with id a4c26156a2ca99f389a40be72f500d6e910a2434e270a77532da488bbeba904f Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.666165 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"e27746ea-410c-45a2-91f5-58633d7c7747","Type":"ContainerStarted","Data":"3d4af5bc68905e57539e64467c20017578833617ba203c9c9d2c32b8e9d9322c"} Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.666809 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"e27746ea-410c-45a2-91f5-58633d7c7747","Type":"ContainerStarted","Data":"0c83ea8a637db8d1cb4bb2df8565e32366cf71162522943ab367ff0fc3432b12"} Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.666827 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"e27746ea-410c-45a2-91f5-58633d7c7747","Type":"ContainerStarted","Data":"a4c26156a2ca99f389a40be72f500d6e910a2434e270a77532da488bbeba904f"} Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.669969 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"dbe58b04-c73c-4292-825f-1e6b466ad54f","Type":"ContainerStarted","Data":"acf32d6e2f919c5e9ee264b5904421b0b72317d31889d6ff084df592e457224c"} Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.670006 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"dbe58b04-c73c-4292-825f-1e6b466ad54f","Type":"ContainerStarted","Data":"e2b4a6e59bf46f86218de4e0fe87152baceec3a79f4dd5da3809c57cacf33d1c"} Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.676941 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"a22fae28-039f-4b62-b889-f53da0ef827b","Type":"ContainerStarted","Data":"2e075cb3f7dd0f5399790c2caa1d403516bf0b0839ef6da3cee24be20b868971"} Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.676985 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"a22fae28-039f-4b62-b889-f53da0ef827b","Type":"ContainerStarted","Data":"545142ab90ccb6fd74836f9de9f78969061b95bb3131524b3b150c0451619564"} Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.696306 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-0" podStartSLOduration=2.61411605 podStartE2EDuration="2.696285336s" podCreationTimestamp="2026-03-09 19:13:42 +0000 UTC" firstStartedPulling="2026-03-09 19:13:43.66090246 +0000 UTC m=+2905.003374858" lastFinishedPulling="2026-03-09 19:13:43.743071746 +0000 UTC m=+2905.085544144" observedRunningTime="2026-03-09 19:13:44.68456401 +0000 UTC m=+2906.027036428" watchObservedRunningTime="2026-03-09 19:13:44.696285336 +0000 UTC m=+2906.038757744" Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.717154 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.525345614 podStartE2EDuration="2.717131498s" podCreationTimestamp="2026-03-09 19:13:42 +0000 UTC" firstStartedPulling="2026-03-09 19:13:43.327819013 +0000 UTC m=+2904.670291411" lastFinishedPulling="2026-03-09 19:13:43.519604897 +0000 UTC m=+2904.862077295" observedRunningTime="2026-03-09 19:13:44.706171102 +0000 UTC m=+2906.048643520" watchObservedRunningTime="2026-03-09 19:13:44.717131498 +0000 UTC m=+2906.059603906" Mar 09 19:13:44 crc kubenswrapper[4750]: I0309 19:13:44.730890 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-2-0" podStartSLOduration=2.517357488 podStartE2EDuration="2.730870229s" podCreationTimestamp="2026-03-09 19:13:42 +0000 UTC" firstStartedPulling="2026-03-09 19:13:43.487913512 +0000 UTC m=+2904.830385930" lastFinishedPulling="2026-03-09 19:13:43.701426273 +0000 UTC m=+2905.043898671" observedRunningTime="2026-03-09 19:13:44.726154512 +0000 UTC m=+2906.068626930" watchObservedRunningTime="2026-03-09 19:13:44.730870229 +0000 UTC m=+2906.073342637" Mar 09 19:13:47 crc kubenswrapper[4750]: I0309 19:13:47.749270 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Mar 09 19:13:47 crc kubenswrapper[4750]: I0309 19:13:47.869218 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:47 crc kubenswrapper[4750]: I0309 19:13:47.913909 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.154249 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.296380 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-scripts\") pod \"adc68d49-e0cb-4adf-abce-56a9eae4f937\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.296496 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgvdt\" (UniqueName: \"kubernetes.io/projected/adc68d49-e0cb-4adf-abce-56a9eae4f937-kube-api-access-jgvdt\") pod \"adc68d49-e0cb-4adf-abce-56a9eae4f937\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.296537 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data-custom\") pod \"adc68d49-e0cb-4adf-abce-56a9eae4f937\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.296618 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adc68d49-e0cb-4adf-abce-56a9eae4f937-etc-machine-id\") pod \"adc68d49-e0cb-4adf-abce-56a9eae4f937\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.296693 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-combined-ca-bundle\") pod \"adc68d49-e0cb-4adf-abce-56a9eae4f937\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.296729 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data\") pod \"adc68d49-e0cb-4adf-abce-56a9eae4f937\" (UID: \"adc68d49-e0cb-4adf-abce-56a9eae4f937\") " Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.296781 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adc68d49-e0cb-4adf-abce-56a9eae4f937-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "adc68d49-e0cb-4adf-abce-56a9eae4f937" (UID: "adc68d49-e0cb-4adf-abce-56a9eae4f937"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.297168 4750 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adc68d49-e0cb-4adf-abce-56a9eae4f937-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.304893 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc68d49-e0cb-4adf-abce-56a9eae4f937-kube-api-access-jgvdt" (OuterVolumeSpecName: "kube-api-access-jgvdt") pod "adc68d49-e0cb-4adf-abce-56a9eae4f937" (UID: "adc68d49-e0cb-4adf-abce-56a9eae4f937"). InnerVolumeSpecName "kube-api-access-jgvdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.305018 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-scripts" (OuterVolumeSpecName: "scripts") pod "adc68d49-e0cb-4adf-abce-56a9eae4f937" (UID: "adc68d49-e0cb-4adf-abce-56a9eae4f937"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.328896 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "adc68d49-e0cb-4adf-abce-56a9eae4f937" (UID: "adc68d49-e0cb-4adf-abce-56a9eae4f937"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.378178 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adc68d49-e0cb-4adf-abce-56a9eae4f937" (UID: "adc68d49-e0cb-4adf-abce-56a9eae4f937"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.399395 4750 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.399425 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgvdt\" (UniqueName: \"kubernetes.io/projected/adc68d49-e0cb-4adf-abce-56a9eae4f937-kube-api-access-jgvdt\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.399436 4750 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.399446 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.417719 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data" (OuterVolumeSpecName: "config-data") pod "adc68d49-e0cb-4adf-abce-56a9eae4f937" (UID: "adc68d49-e0cb-4adf-abce-56a9eae4f937"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.501246 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc68d49-e0cb-4adf-abce-56a9eae4f937-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.727810 4750 generic.go:334] "Generic (PLEG): container finished" podID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerID="76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2" exitCode=0 Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.727854 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"adc68d49-e0cb-4adf-abce-56a9eae4f937","Type":"ContainerDied","Data":"76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2"} Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.727880 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"adc68d49-e0cb-4adf-abce-56a9eae4f937","Type":"ContainerDied","Data":"e62571cb9e4ef6eac1bb15cc9b927d863f8b08226d9f6052bb0a32100ef3a180"} Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.727896 4750 scope.go:117] "RemoveContainer" containerID="65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.728004 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.760791 4750 scope.go:117] "RemoveContainer" containerID="76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.779924 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.814828 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.848821 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 19:13:49 crc kubenswrapper[4750]: E0309 19:13:49.849247 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="probe" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.849263 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="probe" Mar 09 19:13:49 crc kubenswrapper[4750]: E0309 19:13:49.849297 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="cinder-scheduler" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.849303 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="cinder-scheduler" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.849498 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="cinder-scheduler" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.849517 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" containerName="probe" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.850509 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.855666 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.870358 4750 scope.go:117] "RemoveContainer" containerID="65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.870834 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 19:13:49 crc kubenswrapper[4750]: E0309 19:13:49.877943 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9\": container with ID starting with 65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9 not found: ID does not exist" containerID="65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.877983 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9"} err="failed to get container status \"65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9\": rpc error: code = NotFound desc = could not find container \"65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9\": container with ID starting with 65cf56255293b05988f3adfcde1c3c7de1fd79feb04c215f29574cd9bae826d9 not found: ID does not exist" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.878011 4750 scope.go:117] "RemoveContainer" containerID="76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2" Mar 09 19:13:49 crc kubenswrapper[4750]: E0309 19:13:49.878361 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2\": container with ID starting with 76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2 not found: ID does not exist" containerID="76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2" Mar 09 19:13:49 crc kubenswrapper[4750]: I0309 19:13:49.878386 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2"} err="failed to get container status \"76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2\": rpc error: code = NotFound desc = could not find container \"76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2\": container with ID starting with 76ab10ab52271ce6999ac738d7450c2be99029792096ce0688acf665209335f2 not found: ID does not exist" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.017978 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.018282 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psnlj\" (UniqueName: \"kubernetes.io/projected/e2488a7a-65a2-442f-81bf-5c74ce640dc4-kube-api-access-psnlj\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.018350 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-scripts\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.018395 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.018429 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2488a7a-65a2-442f-81bf-5c74ce640dc4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.018446 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-config-data\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.120516 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-scripts\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.120591 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.120753 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2488a7a-65a2-442f-81bf-5c74ce640dc4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.120782 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-config-data\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.120845 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.120898 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psnlj\" (UniqueName: \"kubernetes.io/projected/e2488a7a-65a2-442f-81bf-5c74ce640dc4-kube-api-access-psnlj\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.137832 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2488a7a-65a2-442f-81bf-5c74ce640dc4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.139188 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-scripts\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.140314 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.152080 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-config-data\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.159144 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2488a7a-65a2-442f-81bf-5c74ce640dc4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.179300 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psnlj\" (UniqueName: \"kubernetes.io/projected/e2488a7a-65a2-442f-81bf-5c74ce640dc4-kube-api-access-psnlj\") pod \"cinder-scheduler-0\" (UID: \"e2488a7a-65a2-442f-81bf-5c74ce640dc4\") " pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.188361 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.648268 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 09 19:13:50 crc kubenswrapper[4750]: I0309 19:13:50.759944 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"7e4f5adab4464168db5c82131ffe87d77fc280a87036b80269d90a6786fe6247"} Mar 09 19:13:51 crc kubenswrapper[4750]: I0309 19:13:51.386607 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adc68d49-e0cb-4adf-abce-56a9eae4f937" path="/var/lib/kubelet/pods/adc68d49-e0cb-4adf-abce-56a9eae4f937/volumes" Mar 09 19:13:51 crc kubenswrapper[4750]: I0309 19:13:51.746362 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:13:51 crc kubenswrapper[4750]: I0309 19:13:51.746602 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:13:51 crc kubenswrapper[4750]: I0309 19:13:51.777994 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"2b96b33b66571a888eab27aada0b5b36e08b7d099d629c119b46577e486d5e85"} Mar 09 19:13:52 crc kubenswrapper[4750]: I0309 19:13:52.797945 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"5a9e19c218fcf85f8d772a830045f6ba38220f53b7ad05f7b72325e9cc998fb6"} Mar 09 19:13:52 crc kubenswrapper[4750]: I0309 19:13:52.937129 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Mar 09 19:13:53 crc kubenswrapper[4750]: I0309 19:13:53.002368 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.002344775 podStartE2EDuration="4.002344775s" podCreationTimestamp="2026-03-09 19:13:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 19:13:52.822998976 +0000 UTC m=+2914.165471414" watchObservedRunningTime="2026-03-09 19:13:53.002344775 +0000 UTC m=+2914.344817163" Mar 09 19:13:53 crc kubenswrapper[4750]: I0309 19:13:53.074127 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-0" Mar 09 19:13:53 crc kubenswrapper[4750]: I0309 19:13:53.081114 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-2-0" Mar 09 19:13:54 crc kubenswrapper[4750]: I0309 19:13:54.851437 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="2b96b33b66571a888eab27aada0b5b36e08b7d099d629c119b46577e486d5e85" exitCode=1 Mar 09 19:13:54 crc kubenswrapper[4750]: I0309 19:13:54.851517 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"2b96b33b66571a888eab27aada0b5b36e08b7d099d629c119b46577e486d5e85"} Mar 09 19:13:54 crc kubenswrapper[4750]: I0309 19:13:54.852585 4750 scope.go:117] "RemoveContainer" containerID="2b96b33b66571a888eab27aada0b5b36e08b7d099d629c119b46577e486d5e85" Mar 09 19:13:55 crc kubenswrapper[4750]: I0309 19:13:55.195979 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:13:55 crc kubenswrapper[4750]: I0309 19:13:55.196037 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:13:55 crc kubenswrapper[4750]: I0309 19:13:55.867997 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b"} Mar 09 19:13:57 crc kubenswrapper[4750]: I0309 19:13:57.892442 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b" exitCode=1 Mar 09 19:13:57 crc kubenswrapper[4750]: I0309 19:13:57.893044 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b"} Mar 09 19:13:57 crc kubenswrapper[4750]: I0309 19:13:57.893094 4750 scope.go:117] "RemoveContainer" containerID="2b96b33b66571a888eab27aada0b5b36e08b7d099d629c119b46577e486d5e85" Mar 09 19:13:57 crc kubenswrapper[4750]: I0309 19:13:57.894109 4750 scope.go:117] "RemoveContainer" containerID="9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b" Mar 09 19:13:57 crc kubenswrapper[4750]: E0309 19:13:57.894566 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:13:59 crc kubenswrapper[4750]: I0309 19:13:59.190336 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:13:59 crc kubenswrapper[4750]: I0309 19:13:59.191506 4750 scope.go:117] "RemoveContainer" containerID="9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b" Mar 09 19:13:59 crc kubenswrapper[4750]: E0309 19:13:59.191839 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.142697 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551394-svqlz"] Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.144538 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551394-svqlz" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.147920 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.148425 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.148463 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.154282 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551394-svqlz"] Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.190224 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.190471 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.191358 4750 scope.go:117] "RemoveContainer" containerID="9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b" Mar 09 19:14:00 crc kubenswrapper[4750]: E0309 19:14:00.191616 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.264254 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdmbj\" (UniqueName: \"kubernetes.io/projected/c683fa0a-42f4-4f36-9604-872c91065c80-kube-api-access-gdmbj\") pod \"auto-csr-approver-29551394-svqlz\" (UID: \"c683fa0a-42f4-4f36-9604-872c91065c80\") " pod="openshift-infra/auto-csr-approver-29551394-svqlz" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.366367 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdmbj\" (UniqueName: \"kubernetes.io/projected/c683fa0a-42f4-4f36-9604-872c91065c80-kube-api-access-gdmbj\") pod \"auto-csr-approver-29551394-svqlz\" (UID: \"c683fa0a-42f4-4f36-9604-872c91065c80\") " pod="openshift-infra/auto-csr-approver-29551394-svqlz" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.394277 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdmbj\" (UniqueName: \"kubernetes.io/projected/c683fa0a-42f4-4f36-9604-872c91065c80-kube-api-access-gdmbj\") pod \"auto-csr-approver-29551394-svqlz\" (UID: \"c683fa0a-42f4-4f36-9604-872c91065c80\") " pod="openshift-infra/auto-csr-approver-29551394-svqlz" Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.477043 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551394-svqlz" Mar 09 19:14:00 crc kubenswrapper[4750]: W0309 19:14:00.985724 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc683fa0a_42f4_4f36_9604_872c91065c80.slice/crio-33933911995d860070dbddf45cd4bbcdf5bb9ae8ec12fcea627986080673fcda WatchSource:0}: Error finding container 33933911995d860070dbddf45cd4bbcdf5bb9ae8ec12fcea627986080673fcda: Status 404 returned error can't find the container with id 33933911995d860070dbddf45cd4bbcdf5bb9ae8ec12fcea627986080673fcda Mar 09 19:14:00 crc kubenswrapper[4750]: I0309 19:14:00.988100 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551394-svqlz"] Mar 09 19:14:01 crc kubenswrapper[4750]: I0309 19:14:01.944275 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551394-svqlz" event={"ID":"c683fa0a-42f4-4f36-9604-872c91065c80","Type":"ContainerStarted","Data":"33933911995d860070dbddf45cd4bbcdf5bb9ae8ec12fcea627986080673fcda"} Mar 09 19:14:02 crc kubenswrapper[4750]: I0309 19:14:02.966041 4750 generic.go:334] "Generic (PLEG): container finished" podID="c683fa0a-42f4-4f36-9604-872c91065c80" containerID="c7e67420a73aee4a5304d2e0ae787887c3143abce20047ec675cc0aa54f64c2b" exitCode=0 Mar 09 19:14:02 crc kubenswrapper[4750]: I0309 19:14:02.966095 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551394-svqlz" event={"ID":"c683fa0a-42f4-4f36-9604-872c91065c80","Type":"ContainerDied","Data":"c7e67420a73aee4a5304d2e0ae787887c3143abce20047ec675cc0aa54f64c2b"} Mar 09 19:14:04 crc kubenswrapper[4750]: I0309 19:14:04.393190 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551394-svqlz" Mar 09 19:14:04 crc kubenswrapper[4750]: I0309 19:14:04.474583 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdmbj\" (UniqueName: \"kubernetes.io/projected/c683fa0a-42f4-4f36-9604-872c91065c80-kube-api-access-gdmbj\") pod \"c683fa0a-42f4-4f36-9604-872c91065c80\" (UID: \"c683fa0a-42f4-4f36-9604-872c91065c80\") " Mar 09 19:14:04 crc kubenswrapper[4750]: I0309 19:14:04.480162 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c683fa0a-42f4-4f36-9604-872c91065c80-kube-api-access-gdmbj" (OuterVolumeSpecName: "kube-api-access-gdmbj") pod "c683fa0a-42f4-4f36-9604-872c91065c80" (UID: "c683fa0a-42f4-4f36-9604-872c91065c80"). InnerVolumeSpecName "kube-api-access-gdmbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:14:04 crc kubenswrapper[4750]: I0309 19:14:04.577348 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdmbj\" (UniqueName: \"kubernetes.io/projected/c683fa0a-42f4-4f36-9604-872c91065c80-kube-api-access-gdmbj\") on node \"crc\" DevicePath \"\"" Mar 09 19:14:04 crc kubenswrapper[4750]: I0309 19:14:04.996163 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551394-svqlz" event={"ID":"c683fa0a-42f4-4f36-9604-872c91065c80","Type":"ContainerDied","Data":"33933911995d860070dbddf45cd4bbcdf5bb9ae8ec12fcea627986080673fcda"} Mar 09 19:14:04 crc kubenswrapper[4750]: I0309 19:14:04.996201 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33933911995d860070dbddf45cd4bbcdf5bb9ae8ec12fcea627986080673fcda" Mar 09 19:14:04 crc kubenswrapper[4750]: I0309 19:14:04.996275 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551394-svqlz" Mar 09 19:14:05 crc kubenswrapper[4750]: I0309 19:14:05.487377 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551388-hf8p8"] Mar 09 19:14:05 crc kubenswrapper[4750]: I0309 19:14:05.496662 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551388-hf8p8"] Mar 09 19:14:07 crc kubenswrapper[4750]: I0309 19:14:07.394538 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="160d1e84-3414-4465-8ad7-d142374304aa" path="/var/lib/kubelet/pods/160d1e84-3414-4465-8ad7-d142374304aa/volumes" Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.771800 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ldgdc"] Mar 09 19:14:12 crc kubenswrapper[4750]: E0309 19:14:12.772944 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c683fa0a-42f4-4f36-9604-872c91065c80" containerName="oc" Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.772967 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c683fa0a-42f4-4f36-9604-872c91065c80" containerName="oc" Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.773321 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c683fa0a-42f4-4f36-9604-872c91065c80" containerName="oc" Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.775615 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.802869 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldgdc"] Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.925973 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-catalog-content\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.926219 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj5j6\" (UniqueName: \"kubernetes.io/projected/2e58c0ae-b138-4653-b031-4e446143dc9c-kube-api-access-kj5j6\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:12 crc kubenswrapper[4750]: I0309 19:14:12.926499 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-utilities\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.029906 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-catalog-content\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.030021 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj5j6\" (UniqueName: \"kubernetes.io/projected/2e58c0ae-b138-4653-b031-4e446143dc9c-kube-api-access-kj5j6\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.030050 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-utilities\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.030797 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-utilities\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.031108 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-catalog-content\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.057136 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj5j6\" (UniqueName: \"kubernetes.io/projected/2e58c0ae-b138-4653-b031-4e446143dc9c-kube-api-access-kj5j6\") pod \"redhat-marketplace-ldgdc\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.136198 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:13 crc kubenswrapper[4750]: I0309 19:14:13.614947 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldgdc"] Mar 09 19:14:14 crc kubenswrapper[4750]: I0309 19:14:14.104362 4750 generic.go:334] "Generic (PLEG): container finished" podID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerID="a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d" exitCode=0 Mar 09 19:14:14 crc kubenswrapper[4750]: I0309 19:14:14.104453 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldgdc" event={"ID":"2e58c0ae-b138-4653-b031-4e446143dc9c","Type":"ContainerDied","Data":"a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d"} Mar 09 19:14:14 crc kubenswrapper[4750]: I0309 19:14:14.104669 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldgdc" event={"ID":"2e58c0ae-b138-4653-b031-4e446143dc9c","Type":"ContainerStarted","Data":"ea77154833decd42d7718e3c2dd92842a9c6ea0da74a8a2a93410b6364d143e5"} Mar 09 19:14:15 crc kubenswrapper[4750]: I0309 19:14:15.121897 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldgdc" event={"ID":"2e58c0ae-b138-4653-b031-4e446143dc9c","Type":"ContainerStarted","Data":"f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a"} Mar 09 19:14:15 crc kubenswrapper[4750]: I0309 19:14:15.376757 4750 scope.go:117] "RemoveContainer" containerID="9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b" Mar 09 19:14:16 crc kubenswrapper[4750]: I0309 19:14:16.138759 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0"} Mar 09 19:14:17 crc kubenswrapper[4750]: I0309 19:14:17.150227 4750 generic.go:334] "Generic (PLEG): container finished" podID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerID="f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a" exitCode=0 Mar 09 19:14:17 crc kubenswrapper[4750]: I0309 19:14:17.150439 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldgdc" event={"ID":"2e58c0ae-b138-4653-b031-4e446143dc9c","Type":"ContainerDied","Data":"f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a"} Mar 09 19:14:18 crc kubenswrapper[4750]: I0309 19:14:18.182963 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldgdc" event={"ID":"2e58c0ae-b138-4653-b031-4e446143dc9c","Type":"ContainerStarted","Data":"cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33"} Mar 09 19:14:18 crc kubenswrapper[4750]: I0309 19:14:18.219222 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ldgdc" podStartSLOduration=2.729096496 podStartE2EDuration="6.219192134s" podCreationTimestamp="2026-03-09 19:14:12 +0000 UTC" firstStartedPulling="2026-03-09 19:14:14.108032018 +0000 UTC m=+2935.450504416" lastFinishedPulling="2026-03-09 19:14:17.598127646 +0000 UTC m=+2938.940600054" observedRunningTime="2026-03-09 19:14:18.212158454 +0000 UTC m=+2939.554630862" watchObservedRunningTime="2026-03-09 19:14:18.219192134 +0000 UTC m=+2939.561664552" Mar 09 19:14:19 crc kubenswrapper[4750]: I0309 19:14:19.197758 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0" exitCode=1 Mar 09 19:14:19 crc kubenswrapper[4750]: I0309 19:14:19.197841 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0"} Mar 09 19:14:19 crc kubenswrapper[4750]: I0309 19:14:19.199008 4750 scope.go:117] "RemoveContainer" containerID="9c5672ce427217a13aa3742c571ebf4fa2a3d67c6b0342ec8520188e7774934b" Mar 09 19:14:19 crc kubenswrapper[4750]: I0309 19:14:19.199810 4750 scope.go:117] "RemoveContainer" containerID="2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0" Mar 09 19:14:19 crc kubenswrapper[4750]: E0309 19:14:19.200136 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:14:20 crc kubenswrapper[4750]: I0309 19:14:20.189614 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:20 crc kubenswrapper[4750]: I0309 19:14:20.189957 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:20 crc kubenswrapper[4750]: I0309 19:14:20.189979 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:20 crc kubenswrapper[4750]: I0309 19:14:20.215198 4750 scope.go:117] "RemoveContainer" containerID="2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0" Mar 09 19:14:20 crc kubenswrapper[4750]: E0309 19:14:20.215671 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:14:21 crc kubenswrapper[4750]: I0309 19:14:21.743127 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:14:21 crc kubenswrapper[4750]: I0309 19:14:21.743523 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:14:23 crc kubenswrapper[4750]: I0309 19:14:23.137011 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:23 crc kubenswrapper[4750]: I0309 19:14:23.137929 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:23 crc kubenswrapper[4750]: I0309 19:14:23.212397 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:23 crc kubenswrapper[4750]: I0309 19:14:23.329976 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:23 crc kubenswrapper[4750]: I0309 19:14:23.456120 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldgdc"] Mar 09 19:14:25 crc kubenswrapper[4750]: I0309 19:14:25.279184 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ldgdc" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="registry-server" containerID="cri-o://cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33" gracePeriod=2 Mar 09 19:14:25 crc kubenswrapper[4750]: I0309 19:14:25.910727 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.049400 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-catalog-content\") pod \"2e58c0ae-b138-4653-b031-4e446143dc9c\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.049515 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj5j6\" (UniqueName: \"kubernetes.io/projected/2e58c0ae-b138-4653-b031-4e446143dc9c-kube-api-access-kj5j6\") pod \"2e58c0ae-b138-4653-b031-4e446143dc9c\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.049585 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-utilities\") pod \"2e58c0ae-b138-4653-b031-4e446143dc9c\" (UID: \"2e58c0ae-b138-4653-b031-4e446143dc9c\") " Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.050325 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-utilities" (OuterVolumeSpecName: "utilities") pod "2e58c0ae-b138-4653-b031-4e446143dc9c" (UID: "2e58c0ae-b138-4653-b031-4e446143dc9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.050691 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.055206 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e58c0ae-b138-4653-b031-4e446143dc9c-kube-api-access-kj5j6" (OuterVolumeSpecName: "kube-api-access-kj5j6") pod "2e58c0ae-b138-4653-b031-4e446143dc9c" (UID: "2e58c0ae-b138-4653-b031-4e446143dc9c"). InnerVolumeSpecName "kube-api-access-kj5j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.077177 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e58c0ae-b138-4653-b031-4e446143dc9c" (UID: "2e58c0ae-b138-4653-b031-4e446143dc9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.152863 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e58c0ae-b138-4653-b031-4e446143dc9c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.152904 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj5j6\" (UniqueName: \"kubernetes.io/projected/2e58c0ae-b138-4653-b031-4e446143dc9c-kube-api-access-kj5j6\") on node \"crc\" DevicePath \"\"" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.298209 4750 generic.go:334] "Generic (PLEG): container finished" podID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerID="cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33" exitCode=0 Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.298260 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldgdc" event={"ID":"2e58c0ae-b138-4653-b031-4e446143dc9c","Type":"ContainerDied","Data":"cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33"} Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.298324 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldgdc" event={"ID":"2e58c0ae-b138-4653-b031-4e446143dc9c","Type":"ContainerDied","Data":"ea77154833decd42d7718e3c2dd92842a9c6ea0da74a8a2a93410b6364d143e5"} Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.298320 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldgdc" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.298346 4750 scope.go:117] "RemoveContainer" containerID="cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.361274 4750 scope.go:117] "RemoveContainer" containerID="f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.365657 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldgdc"] Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.378214 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldgdc"] Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.399126 4750 scope.go:117] "RemoveContainer" containerID="a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.477647 4750 scope.go:117] "RemoveContainer" containerID="cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33" Mar 09 19:14:26 crc kubenswrapper[4750]: E0309 19:14:26.478450 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33\": container with ID starting with cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33 not found: ID does not exist" containerID="cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.478510 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33"} err="failed to get container status \"cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33\": rpc error: code = NotFound desc = could not find container \"cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33\": container with ID starting with cddce46ba22becfcae4284ba48145e1a23690ef0633bd3b6125474dba752fb33 not found: ID does not exist" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.478546 4750 scope.go:117] "RemoveContainer" containerID="f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a" Mar 09 19:14:26 crc kubenswrapper[4750]: E0309 19:14:26.479082 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a\": container with ID starting with f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a not found: ID does not exist" containerID="f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.479198 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a"} err="failed to get container status \"f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a\": rpc error: code = NotFound desc = could not find container \"f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a\": container with ID starting with f271d511deea3d56e26d928ad68ee1dd8731f4eac679e147c7b13e347f39726a not found: ID does not exist" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.479291 4750 scope.go:117] "RemoveContainer" containerID="a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d" Mar 09 19:14:26 crc kubenswrapper[4750]: E0309 19:14:26.479771 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d\": container with ID starting with a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d not found: ID does not exist" containerID="a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d" Mar 09 19:14:26 crc kubenswrapper[4750]: I0309 19:14:26.479807 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d"} err="failed to get container status \"a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d\": rpc error: code = NotFound desc = could not find container \"a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d\": container with ID starting with a78ee42995c84895a6b0d5b1cdd479b5d2d971b2c1b952efed303a53d8cf2c8d not found: ID does not exist" Mar 09 19:14:27 crc kubenswrapper[4750]: I0309 19:14:27.394400 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" path="/var/lib/kubelet/pods/2e58c0ae-b138-4653-b031-4e446143dc9c/volumes" Mar 09 19:14:34 crc kubenswrapper[4750]: I0309 19:14:34.402367 4750 scope.go:117] "RemoveContainer" containerID="2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0" Mar 09 19:14:34 crc kubenswrapper[4750]: E0309 19:14:34.403165 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:14:45 crc kubenswrapper[4750]: I0309 19:14:45.922087 4750 scope.go:117] "RemoveContainer" containerID="fbebdef272d543507373f4753a288dccf1a57dad80d276ede115a4b663cd465d" Mar 09 19:14:48 crc kubenswrapper[4750]: I0309 19:14:48.374047 4750 scope.go:117] "RemoveContainer" containerID="2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0" Mar 09 19:14:49 crc kubenswrapper[4750]: I0309 19:14:49.613571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d"} Mar 09 19:14:50 crc kubenswrapper[4750]: I0309 19:14:50.189909 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.647751 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" exitCode=1 Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.647851 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d"} Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.648266 4750 scope.go:117] "RemoveContainer" containerID="2c860d58896cd696a8e9820f1a7a94f0f4bec3ee32ba266fc6368ef8252505e0" Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.649069 4750 scope.go:117] "RemoveContainer" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" Mar 09 19:14:51 crc kubenswrapper[4750]: E0309 19:14:51.649739 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.745540 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.745685 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.745789 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.748420 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f78f5d76d0e8b5025f0101c0000bfefe714e5ba77c6b39281d9459bc8318a103"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:14:51 crc kubenswrapper[4750]: I0309 19:14:51.748535 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://f78f5d76d0e8b5025f0101c0000bfefe714e5ba77c6b39281d9459bc8318a103" gracePeriod=600 Mar 09 19:14:52 crc kubenswrapper[4750]: I0309 19:14:52.668889 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="f78f5d76d0e8b5025f0101c0000bfefe714e5ba77c6b39281d9459bc8318a103" exitCode=0 Mar 09 19:14:52 crc kubenswrapper[4750]: I0309 19:14:52.668961 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"f78f5d76d0e8b5025f0101c0000bfefe714e5ba77c6b39281d9459bc8318a103"} Mar 09 19:14:52 crc kubenswrapper[4750]: I0309 19:14:52.669283 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d"} Mar 09 19:14:52 crc kubenswrapper[4750]: I0309 19:14:52.669319 4750 scope.go:117] "RemoveContainer" containerID="464357c4dbad94a4f7129c3b906a8b2efe3e503d4408eb1e35d26dc4c60f97db" Mar 09 19:14:53 crc kubenswrapper[4750]: I0309 19:14:53.189939 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:53 crc kubenswrapper[4750]: I0309 19:14:53.191076 4750 scope.go:117] "RemoveContainer" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" Mar 09 19:14:53 crc kubenswrapper[4750]: E0309 19:14:53.191667 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:14:55 crc kubenswrapper[4750]: I0309 19:14:55.189439 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:14:55 crc kubenswrapper[4750]: I0309 19:14:55.191737 4750 scope.go:117] "RemoveContainer" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" Mar 09 19:14:55 crc kubenswrapper[4750]: E0309 19:14:55.192274 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.192254 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f"] Mar 09 19:15:00 crc kubenswrapper[4750]: E0309 19:15:00.193879 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="registry-server" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.193917 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="registry-server" Mar 09 19:15:00 crc kubenswrapper[4750]: E0309 19:15:00.193984 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="extract-content" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.194004 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="extract-content" Mar 09 19:15:00 crc kubenswrapper[4750]: E0309 19:15:00.194036 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="extract-utilities" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.194055 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="extract-utilities" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.194593 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e58c0ae-b138-4653-b031-4e446143dc9c" containerName="registry-server" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.196375 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.199893 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.200001 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.210479 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f"] Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.368764 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d42a7252-53c5-411e-8b76-20926501ab47-config-volume\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.368824 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc4dp\" (UniqueName: \"kubernetes.io/projected/d42a7252-53c5-411e-8b76-20926501ab47-kube-api-access-hc4dp\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.368894 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d42a7252-53c5-411e-8b76-20926501ab47-secret-volume\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.470310 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d42a7252-53c5-411e-8b76-20926501ab47-config-volume\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.470358 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc4dp\" (UniqueName: \"kubernetes.io/projected/d42a7252-53c5-411e-8b76-20926501ab47-kube-api-access-hc4dp\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.470381 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d42a7252-53c5-411e-8b76-20926501ab47-secret-volume\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.472296 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d42a7252-53c5-411e-8b76-20926501ab47-config-volume\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.477718 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d42a7252-53c5-411e-8b76-20926501ab47-secret-volume\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.487768 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc4dp\" (UniqueName: \"kubernetes.io/projected/d42a7252-53c5-411e-8b76-20926501ab47-kube-api-access-hc4dp\") pod \"collect-profiles-29551395-2fp5f\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:00 crc kubenswrapper[4750]: I0309 19:15:00.527222 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:01 crc kubenswrapper[4750]: I0309 19:15:01.062842 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f"] Mar 09 19:15:01 crc kubenswrapper[4750]: I0309 19:15:01.781573 4750 generic.go:334] "Generic (PLEG): container finished" podID="d42a7252-53c5-411e-8b76-20926501ab47" containerID="b24d417c6f76a416f23654c1ddf46b1d0b72a69749ee6bad769126c64e1a69c6" exitCode=0 Mar 09 19:15:01 crc kubenswrapper[4750]: I0309 19:15:01.781677 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" event={"ID":"d42a7252-53c5-411e-8b76-20926501ab47","Type":"ContainerDied","Data":"b24d417c6f76a416f23654c1ddf46b1d0b72a69749ee6bad769126c64e1a69c6"} Mar 09 19:15:01 crc kubenswrapper[4750]: I0309 19:15:01.782207 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" event={"ID":"d42a7252-53c5-411e-8b76-20926501ab47","Type":"ContainerStarted","Data":"aa1f3080ee74554c9f4cd958489bb5a7dc4e6479004ad0ddb03be61d026785d3"} Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.229802 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.355078 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc4dp\" (UniqueName: \"kubernetes.io/projected/d42a7252-53c5-411e-8b76-20926501ab47-kube-api-access-hc4dp\") pod \"d42a7252-53c5-411e-8b76-20926501ab47\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.355381 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d42a7252-53c5-411e-8b76-20926501ab47-secret-volume\") pod \"d42a7252-53c5-411e-8b76-20926501ab47\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.355451 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d42a7252-53c5-411e-8b76-20926501ab47-config-volume\") pod \"d42a7252-53c5-411e-8b76-20926501ab47\" (UID: \"d42a7252-53c5-411e-8b76-20926501ab47\") " Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.356085 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d42a7252-53c5-411e-8b76-20926501ab47-config-volume" (OuterVolumeSpecName: "config-volume") pod "d42a7252-53c5-411e-8b76-20926501ab47" (UID: "d42a7252-53c5-411e-8b76-20926501ab47"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.361142 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d42a7252-53c5-411e-8b76-20926501ab47-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d42a7252-53c5-411e-8b76-20926501ab47" (UID: "d42a7252-53c5-411e-8b76-20926501ab47"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.369873 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d42a7252-53c5-411e-8b76-20926501ab47-kube-api-access-hc4dp" (OuterVolumeSpecName: "kube-api-access-hc4dp") pod "d42a7252-53c5-411e-8b76-20926501ab47" (UID: "d42a7252-53c5-411e-8b76-20926501ab47"). InnerVolumeSpecName "kube-api-access-hc4dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.458257 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d42a7252-53c5-411e-8b76-20926501ab47-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.458454 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d42a7252-53c5-411e-8b76-20926501ab47-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.458537 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc4dp\" (UniqueName: \"kubernetes.io/projected/d42a7252-53c5-411e-8b76-20926501ab47-kube-api-access-hc4dp\") on node \"crc\" DevicePath \"\"" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.810253 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" event={"ID":"d42a7252-53c5-411e-8b76-20926501ab47","Type":"ContainerDied","Data":"aa1f3080ee74554c9f4cd958489bb5a7dc4e6479004ad0ddb03be61d026785d3"} Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.810293 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa1f3080ee74554c9f4cd958489bb5a7dc4e6479004ad0ddb03be61d026785d3" Mar 09 19:15:03 crc kubenswrapper[4750]: I0309 19:15:03.810542 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f" Mar 09 19:15:04 crc kubenswrapper[4750]: I0309 19:15:04.331336 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl"] Mar 09 19:15:04 crc kubenswrapper[4750]: I0309 19:15:04.339884 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551350-2zzrl"] Mar 09 19:15:05 crc kubenswrapper[4750]: I0309 19:15:05.397734 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1069377-fcc6-4478-92cb-a93487954a97" path="/var/lib/kubelet/pods/b1069377-fcc6-4478-92cb-a93487954a97/volumes" Mar 09 19:15:07 crc kubenswrapper[4750]: I0309 19:15:07.374565 4750 scope.go:117] "RemoveContainer" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" Mar 09 19:15:07 crc kubenswrapper[4750]: E0309 19:15:07.375523 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:15:18 crc kubenswrapper[4750]: I0309 19:15:18.373667 4750 scope.go:117] "RemoveContainer" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" Mar 09 19:15:18 crc kubenswrapper[4750]: E0309 19:15:18.375393 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:15:33 crc kubenswrapper[4750]: I0309 19:15:33.374422 4750 scope.go:117] "RemoveContainer" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" Mar 09 19:15:34 crc kubenswrapper[4750]: I0309 19:15:34.233259 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a"} Mar 09 19:15:35 crc kubenswrapper[4750]: I0309 19:15:35.190059 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:15:36 crc kubenswrapper[4750]: I0309 19:15:36.266919 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" exitCode=1 Mar 09 19:15:36 crc kubenswrapper[4750]: I0309 19:15:36.267004 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a"} Mar 09 19:15:36 crc kubenswrapper[4750]: I0309 19:15:36.267355 4750 scope.go:117] "RemoveContainer" containerID="4dae0ed775dc3a43c4e6df4bf511a398b98a4bae3c935c0fbc003721b5aceb9d" Mar 09 19:15:36 crc kubenswrapper[4750]: I0309 19:15:36.268644 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:15:36 crc kubenswrapper[4750]: E0309 19:15:36.269330 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:15:38 crc kubenswrapper[4750]: I0309 19:15:38.189755 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:15:38 crc kubenswrapper[4750]: I0309 19:15:38.191404 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:15:38 crc kubenswrapper[4750]: E0309 19:15:38.191991 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:15:40 crc kubenswrapper[4750]: I0309 19:15:40.189621 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:15:40 crc kubenswrapper[4750]: I0309 19:15:40.191120 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:15:40 crc kubenswrapper[4750]: E0309 19:15:40.191515 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:15:46 crc kubenswrapper[4750]: I0309 19:15:46.076614 4750 scope.go:117] "RemoveContainer" containerID="1466913896df81ea73a2eee20a325bdc8813bed7a46c93fee0925d84030bd21a" Mar 09 19:15:53 crc kubenswrapper[4750]: I0309 19:15:53.373489 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:15:53 crc kubenswrapper[4750]: E0309 19:15:53.375416 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.207949 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551396-nqxcx"] Mar 09 19:16:00 crc kubenswrapper[4750]: E0309 19:16:00.209277 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d42a7252-53c5-411e-8b76-20926501ab47" containerName="collect-profiles" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.209307 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d42a7252-53c5-411e-8b76-20926501ab47" containerName="collect-profiles" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.209838 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d42a7252-53c5-411e-8b76-20926501ab47" containerName="collect-profiles" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.210707 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.215651 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.216805 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.220411 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551396-nqxcx"] Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.220485 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.323648 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpgqh\" (UniqueName: \"kubernetes.io/projected/ff284095-d8e1-4ad3-9b21-a3ea610e448c-kube-api-access-vpgqh\") pod \"auto-csr-approver-29551396-nqxcx\" (UID: \"ff284095-d8e1-4ad3-9b21-a3ea610e448c\") " pod="openshift-infra/auto-csr-approver-29551396-nqxcx" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.425549 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpgqh\" (UniqueName: \"kubernetes.io/projected/ff284095-d8e1-4ad3-9b21-a3ea610e448c-kube-api-access-vpgqh\") pod \"auto-csr-approver-29551396-nqxcx\" (UID: \"ff284095-d8e1-4ad3-9b21-a3ea610e448c\") " pod="openshift-infra/auto-csr-approver-29551396-nqxcx" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.447473 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpgqh\" (UniqueName: \"kubernetes.io/projected/ff284095-d8e1-4ad3-9b21-a3ea610e448c-kube-api-access-vpgqh\") pod \"auto-csr-approver-29551396-nqxcx\" (UID: \"ff284095-d8e1-4ad3-9b21-a3ea610e448c\") " pod="openshift-infra/auto-csr-approver-29551396-nqxcx" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.544110 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" Mar 09 19:16:00 crc kubenswrapper[4750]: I0309 19:16:00.850357 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551396-nqxcx"] Mar 09 19:16:01 crc kubenswrapper[4750]: I0309 19:16:01.583048 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" event={"ID":"ff284095-d8e1-4ad3-9b21-a3ea610e448c","Type":"ContainerStarted","Data":"5d9a32657bc82044ed80238730bac279bc0fcd02147a57879500e41ebd368bf9"} Mar 09 19:16:02 crc kubenswrapper[4750]: I0309 19:16:02.594196 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" event={"ID":"ff284095-d8e1-4ad3-9b21-a3ea610e448c","Type":"ContainerStarted","Data":"43eafd751ca9b9e68f2a74e9a5ec15856ae2d361796c923b4a721acfc9184fb7"} Mar 09 19:16:02 crc kubenswrapper[4750]: I0309 19:16:02.617886 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" podStartSLOduration=1.5346135520000002 podStartE2EDuration="2.617862862s" podCreationTimestamp="2026-03-09 19:16:00 +0000 UTC" firstStartedPulling="2026-03-09 19:16:00.856899871 +0000 UTC m=+3042.199372269" lastFinishedPulling="2026-03-09 19:16:01.940149181 +0000 UTC m=+3043.282621579" observedRunningTime="2026-03-09 19:16:02.611402457 +0000 UTC m=+3043.953874865" watchObservedRunningTime="2026-03-09 19:16:02.617862862 +0000 UTC m=+3043.960335270" Mar 09 19:16:03 crc kubenswrapper[4750]: I0309 19:16:03.607779 4750 generic.go:334] "Generic (PLEG): container finished" podID="ff284095-d8e1-4ad3-9b21-a3ea610e448c" containerID="43eafd751ca9b9e68f2a74e9a5ec15856ae2d361796c923b4a721acfc9184fb7" exitCode=0 Mar 09 19:16:03 crc kubenswrapper[4750]: I0309 19:16:03.607864 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" event={"ID":"ff284095-d8e1-4ad3-9b21-a3ea610e448c","Type":"ContainerDied","Data":"43eafd751ca9b9e68f2a74e9a5ec15856ae2d361796c923b4a721acfc9184fb7"} Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.109961 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.248094 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpgqh\" (UniqueName: \"kubernetes.io/projected/ff284095-d8e1-4ad3-9b21-a3ea610e448c-kube-api-access-vpgqh\") pod \"ff284095-d8e1-4ad3-9b21-a3ea610e448c\" (UID: \"ff284095-d8e1-4ad3-9b21-a3ea610e448c\") " Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.256997 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff284095-d8e1-4ad3-9b21-a3ea610e448c-kube-api-access-vpgqh" (OuterVolumeSpecName: "kube-api-access-vpgqh") pod "ff284095-d8e1-4ad3-9b21-a3ea610e448c" (UID: "ff284095-d8e1-4ad3-9b21-a3ea610e448c"). InnerVolumeSpecName "kube-api-access-vpgqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.352251 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpgqh\" (UniqueName: \"kubernetes.io/projected/ff284095-d8e1-4ad3-9b21-a3ea610e448c-kube-api-access-vpgqh\") on node \"crc\" DevicePath \"\"" Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.642690 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" event={"ID":"ff284095-d8e1-4ad3-9b21-a3ea610e448c","Type":"ContainerDied","Data":"5d9a32657bc82044ed80238730bac279bc0fcd02147a57879500e41ebd368bf9"} Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.642753 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d9a32657bc82044ed80238730bac279bc0fcd02147a57879500e41ebd368bf9" Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.642838 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551396-nqxcx" Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.711551 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551390-xzz7r"] Mar 09 19:16:05 crc kubenswrapper[4750]: I0309 19:16:05.725150 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551390-xzz7r"] Mar 09 19:16:07 crc kubenswrapper[4750]: I0309 19:16:07.391389 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87d71783-700a-40fd-96f1-ad7f861ccf38" path="/var/lib/kubelet/pods/87d71783-700a-40fd-96f1-ad7f861ccf38/volumes" Mar 09 19:16:08 crc kubenswrapper[4750]: I0309 19:16:08.374669 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:16:08 crc kubenswrapper[4750]: E0309 19:16:08.375216 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:16:19 crc kubenswrapper[4750]: I0309 19:16:19.388359 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:16:19 crc kubenswrapper[4750]: E0309 19:16:19.389138 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:16:31 crc kubenswrapper[4750]: I0309 19:16:31.373870 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:16:31 crc kubenswrapper[4750]: E0309 19:16:31.374666 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:16:45 crc kubenswrapper[4750]: I0309 19:16:45.373853 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:16:45 crc kubenswrapper[4750]: E0309 19:16:45.374949 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:16:46 crc kubenswrapper[4750]: I0309 19:16:46.156607 4750 scope.go:117] "RemoveContainer" containerID="07ea2e62d5b33a075bbc264fe9ae81afb8e75d7c1e339f7f970ac954b33e7f4c" Mar 09 19:16:57 crc kubenswrapper[4750]: I0309 19:16:57.373651 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:16:58 crc kubenswrapper[4750]: I0309 19:16:58.317355 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f"} Mar 09 19:17:00 crc kubenswrapper[4750]: I0309 19:17:00.189483 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:17:00 crc kubenswrapper[4750]: I0309 19:17:00.341877 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" exitCode=1 Mar 09 19:17:00 crc kubenswrapper[4750]: I0309 19:17:00.341978 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f"} Mar 09 19:17:00 crc kubenswrapper[4750]: I0309 19:17:00.342269 4750 scope.go:117] "RemoveContainer" containerID="c5082f08899d4bd8858f12c599be4202d1b65696c45ef943b46475f6e163e59a" Mar 09 19:17:00 crc kubenswrapper[4750]: I0309 19:17:00.344300 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:17:00 crc kubenswrapper[4750]: E0309 19:17:00.347177 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:17:02 crc kubenswrapper[4750]: I0309 19:17:02.189932 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:17:02 crc kubenswrapper[4750]: I0309 19:17:02.190534 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:17:02 crc kubenswrapper[4750]: E0309 19:17:02.190977 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:17:05 crc kubenswrapper[4750]: I0309 19:17:05.190080 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:17:05 crc kubenswrapper[4750]: I0309 19:17:05.191096 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:17:05 crc kubenswrapper[4750]: E0309 19:17:05.191499 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:17:18 crc kubenswrapper[4750]: I0309 19:17:18.373714 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:17:18 crc kubenswrapper[4750]: E0309 19:17:18.374752 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:17:21 crc kubenswrapper[4750]: I0309 19:17:21.743282 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:17:21 crc kubenswrapper[4750]: I0309 19:17:21.743874 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:17:33 crc kubenswrapper[4750]: I0309 19:17:33.374199 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:17:33 crc kubenswrapper[4750]: E0309 19:17:33.375195 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:17:44 crc kubenswrapper[4750]: I0309 19:17:44.373453 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:17:44 crc kubenswrapper[4750]: E0309 19:17:44.375489 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:17:51 crc kubenswrapper[4750]: I0309 19:17:51.743348 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:17:51 crc kubenswrapper[4750]: I0309 19:17:51.743912 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:17:57 crc kubenswrapper[4750]: I0309 19:17:57.373758 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:17:57 crc kubenswrapper[4750]: E0309 19:17:57.374801 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.173029 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551398-znfd4"] Mar 09 19:18:00 crc kubenswrapper[4750]: E0309 19:18:00.174372 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff284095-d8e1-4ad3-9b21-a3ea610e448c" containerName="oc" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.174394 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff284095-d8e1-4ad3-9b21-a3ea610e448c" containerName="oc" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.176164 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff284095-d8e1-4ad3-9b21-a3ea610e448c" containerName="oc" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.177071 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551398-znfd4" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.182030 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.182044 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.182378 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.188357 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551398-znfd4"] Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.363994 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t486\" (UniqueName: \"kubernetes.io/projected/4afddec7-7283-410f-8707-2f4112be0f91-kube-api-access-9t486\") pod \"auto-csr-approver-29551398-znfd4\" (UID: \"4afddec7-7283-410f-8707-2f4112be0f91\") " pod="openshift-infra/auto-csr-approver-29551398-znfd4" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.466323 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t486\" (UniqueName: \"kubernetes.io/projected/4afddec7-7283-410f-8707-2f4112be0f91-kube-api-access-9t486\") pod \"auto-csr-approver-29551398-znfd4\" (UID: \"4afddec7-7283-410f-8707-2f4112be0f91\") " pod="openshift-infra/auto-csr-approver-29551398-znfd4" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.490813 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t486\" (UniqueName: \"kubernetes.io/projected/4afddec7-7283-410f-8707-2f4112be0f91-kube-api-access-9t486\") pod \"auto-csr-approver-29551398-znfd4\" (UID: \"4afddec7-7283-410f-8707-2f4112be0f91\") " pod="openshift-infra/auto-csr-approver-29551398-znfd4" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.505256 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551398-znfd4" Mar 09 19:18:00 crc kubenswrapper[4750]: I0309 19:18:00.981564 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551398-znfd4"] Mar 09 19:18:01 crc kubenswrapper[4750]: I0309 19:18:01.086137 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551398-znfd4" event={"ID":"4afddec7-7283-410f-8707-2f4112be0f91","Type":"ContainerStarted","Data":"5637f3e51b589a21ce9b4e2763748e9078caf14dc85dd16d5b0f1aff4668d7ee"} Mar 09 19:18:03 crc kubenswrapper[4750]: I0309 19:18:03.112907 4750 generic.go:334] "Generic (PLEG): container finished" podID="4afddec7-7283-410f-8707-2f4112be0f91" containerID="de48542d81c0694d3c1764638d52e91e4bc6e50c4ae371f202fc275da0becacb" exitCode=0 Mar 09 19:18:03 crc kubenswrapper[4750]: I0309 19:18:03.113021 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551398-znfd4" event={"ID":"4afddec7-7283-410f-8707-2f4112be0f91","Type":"ContainerDied","Data":"de48542d81c0694d3c1764638d52e91e4bc6e50c4ae371f202fc275da0becacb"} Mar 09 19:18:04 crc kubenswrapper[4750]: I0309 19:18:04.576910 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551398-znfd4" Mar 09 19:18:04 crc kubenswrapper[4750]: I0309 19:18:04.674477 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t486\" (UniqueName: \"kubernetes.io/projected/4afddec7-7283-410f-8707-2f4112be0f91-kube-api-access-9t486\") pod \"4afddec7-7283-410f-8707-2f4112be0f91\" (UID: \"4afddec7-7283-410f-8707-2f4112be0f91\") " Mar 09 19:18:04 crc kubenswrapper[4750]: I0309 19:18:04.685825 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4afddec7-7283-410f-8707-2f4112be0f91-kube-api-access-9t486" (OuterVolumeSpecName: "kube-api-access-9t486") pod "4afddec7-7283-410f-8707-2f4112be0f91" (UID: "4afddec7-7283-410f-8707-2f4112be0f91"). InnerVolumeSpecName "kube-api-access-9t486". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:18:04 crc kubenswrapper[4750]: I0309 19:18:04.776796 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t486\" (UniqueName: \"kubernetes.io/projected/4afddec7-7283-410f-8707-2f4112be0f91-kube-api-access-9t486\") on node \"crc\" DevicePath \"\"" Mar 09 19:18:05 crc kubenswrapper[4750]: I0309 19:18:05.140729 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551398-znfd4" event={"ID":"4afddec7-7283-410f-8707-2f4112be0f91","Type":"ContainerDied","Data":"5637f3e51b589a21ce9b4e2763748e9078caf14dc85dd16d5b0f1aff4668d7ee"} Mar 09 19:18:05 crc kubenswrapper[4750]: I0309 19:18:05.140788 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5637f3e51b589a21ce9b4e2763748e9078caf14dc85dd16d5b0f1aff4668d7ee" Mar 09 19:18:05 crc kubenswrapper[4750]: I0309 19:18:05.140862 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551398-znfd4" Mar 09 19:18:05 crc kubenswrapper[4750]: I0309 19:18:05.682010 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551392-dqmh4"] Mar 09 19:18:05 crc kubenswrapper[4750]: I0309 19:18:05.696420 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551392-dqmh4"] Mar 09 19:18:07 crc kubenswrapper[4750]: I0309 19:18:07.395830 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1527157-9fda-4756-b0aa-8fa60c57b85f" path="/var/lib/kubelet/pods/b1527157-9fda-4756-b0aa-8fa60c57b85f/volumes" Mar 09 19:18:10 crc kubenswrapper[4750]: I0309 19:18:10.374134 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:18:10 crc kubenswrapper[4750]: E0309 19:18:10.375271 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:18:21 crc kubenswrapper[4750]: I0309 19:18:21.743593 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:18:21 crc kubenswrapper[4750]: I0309 19:18:21.744459 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:18:21 crc kubenswrapper[4750]: I0309 19:18:21.744523 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:18:21 crc kubenswrapper[4750]: I0309 19:18:21.745366 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:18:21 crc kubenswrapper[4750]: I0309 19:18:21.745459 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" gracePeriod=600 Mar 09 19:18:21 crc kubenswrapper[4750]: E0309 19:18:21.883790 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:18:22 crc kubenswrapper[4750]: I0309 19:18:22.362303 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" exitCode=0 Mar 09 19:18:22 crc kubenswrapper[4750]: I0309 19:18:22.362356 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d"} Mar 09 19:18:22 crc kubenswrapper[4750]: I0309 19:18:22.362396 4750 scope.go:117] "RemoveContainer" containerID="f78f5d76d0e8b5025f0101c0000bfefe714e5ba77c6b39281d9459bc8318a103" Mar 09 19:18:22 crc kubenswrapper[4750]: I0309 19:18:22.365498 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:18:22 crc kubenswrapper[4750]: E0309 19:18:22.369718 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:18:25 crc kubenswrapper[4750]: I0309 19:18:25.374502 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:18:25 crc kubenswrapper[4750]: E0309 19:18:25.376054 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:18:34 crc kubenswrapper[4750]: I0309 19:18:34.374175 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:18:34 crc kubenswrapper[4750]: E0309 19:18:34.375216 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:18:36 crc kubenswrapper[4750]: I0309 19:18:36.375126 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:18:36 crc kubenswrapper[4750]: E0309 19:18:36.377697 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:18:46 crc kubenswrapper[4750]: I0309 19:18:46.287155 4750 scope.go:117] "RemoveContainer" containerID="e688510fa33c4594f5b624031c2e09eaf5ca3723129ce99e855218e66f924a77" Mar 09 19:18:47 crc kubenswrapper[4750]: I0309 19:18:47.374184 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:18:47 crc kubenswrapper[4750]: E0309 19:18:47.376439 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:18:49 crc kubenswrapper[4750]: I0309 19:18:49.373212 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:18:49 crc kubenswrapper[4750]: E0309 19:18:49.373860 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:18:59 crc kubenswrapper[4750]: I0309 19:18:59.400683 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:18:59 crc kubenswrapper[4750]: E0309 19:18:59.401600 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:19:00 crc kubenswrapper[4750]: I0309 19:19:00.374795 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:19:00 crc kubenswrapper[4750]: E0309 19:19:00.375332 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:19:10 crc kubenswrapper[4750]: I0309 19:19:10.373842 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:19:10 crc kubenswrapper[4750]: E0309 19:19:10.374977 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:19:12 crc kubenswrapper[4750]: I0309 19:19:12.373400 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:19:12 crc kubenswrapper[4750]: E0309 19:19:12.373965 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:19:23 crc kubenswrapper[4750]: I0309 19:19:23.374276 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:19:23 crc kubenswrapper[4750]: E0309 19:19:23.375251 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:19:25 crc kubenswrapper[4750]: I0309 19:19:25.373928 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:19:25 crc kubenswrapper[4750]: E0309 19:19:25.374743 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:19:37 crc kubenswrapper[4750]: I0309 19:19:37.373995 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:19:37 crc kubenswrapper[4750]: E0309 19:19:37.375619 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:19:38 crc kubenswrapper[4750]: I0309 19:19:38.373951 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:19:38 crc kubenswrapper[4750]: E0309 19:19:38.374261 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:19:50 crc kubenswrapper[4750]: I0309 19:19:50.374087 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:19:51 crc kubenswrapper[4750]: I0309 19:19:51.504284 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5"} Mar 09 19:19:53 crc kubenswrapper[4750]: I0309 19:19:53.373496 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:19:53 crc kubenswrapper[4750]: E0309 19:19:53.374764 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:19:53 crc kubenswrapper[4750]: I0309 19:19:53.541081 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" exitCode=1 Mar 09 19:19:53 crc kubenswrapper[4750]: I0309 19:19:53.541142 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5"} Mar 09 19:19:53 crc kubenswrapper[4750]: I0309 19:19:53.541192 4750 scope.go:117] "RemoveContainer" containerID="47b03f50e03f8a77c42c7b07fac72fcadf9f81407f8a7fa4b50c29c2cbfc813f" Mar 09 19:19:53 crc kubenswrapper[4750]: I0309 19:19:53.542118 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:19:53 crc kubenswrapper[4750]: E0309 19:19:53.542499 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:19:55 crc kubenswrapper[4750]: I0309 19:19:55.190374 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:19:55 crc kubenswrapper[4750]: I0309 19:19:55.190861 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:19:55 crc kubenswrapper[4750]: I0309 19:19:55.192093 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:19:55 crc kubenswrapper[4750]: E0309 19:19:55.192620 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:19:56 crc kubenswrapper[4750]: I0309 19:19:56.190296 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:19:56 crc kubenswrapper[4750]: I0309 19:19:56.191907 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:19:56 crc kubenswrapper[4750]: E0309 19:19:56.192528 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.227144 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551400-2jbpn"] Mar 09 19:20:00 crc kubenswrapper[4750]: E0309 19:20:00.228408 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4afddec7-7283-410f-8707-2f4112be0f91" containerName="oc" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.228445 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4afddec7-7283-410f-8707-2f4112be0f91" containerName="oc" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.229030 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4afddec7-7283-410f-8707-2f4112be0f91" containerName="oc" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.230576 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551400-2jbpn" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.234919 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.235866 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.235933 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.248173 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551400-2jbpn"] Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.279297 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdvgb\" (UniqueName: \"kubernetes.io/projected/78e39153-00de-4622-a05a-1b92ecac2777-kube-api-access-sdvgb\") pod \"auto-csr-approver-29551400-2jbpn\" (UID: \"78e39153-00de-4622-a05a-1b92ecac2777\") " pod="openshift-infra/auto-csr-approver-29551400-2jbpn" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.381202 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdvgb\" (UniqueName: \"kubernetes.io/projected/78e39153-00de-4622-a05a-1b92ecac2777-kube-api-access-sdvgb\") pod \"auto-csr-approver-29551400-2jbpn\" (UID: \"78e39153-00de-4622-a05a-1b92ecac2777\") " pod="openshift-infra/auto-csr-approver-29551400-2jbpn" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.423520 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdvgb\" (UniqueName: \"kubernetes.io/projected/78e39153-00de-4622-a05a-1b92ecac2777-kube-api-access-sdvgb\") pod \"auto-csr-approver-29551400-2jbpn\" (UID: \"78e39153-00de-4622-a05a-1b92ecac2777\") " pod="openshift-infra/auto-csr-approver-29551400-2jbpn" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.564921 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551400-2jbpn" Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.932358 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551400-2jbpn"] Mar 09 19:20:00 crc kubenswrapper[4750]: I0309 19:20:00.933284 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:20:01 crc kubenswrapper[4750]: I0309 19:20:01.651683 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551400-2jbpn" event={"ID":"78e39153-00de-4622-a05a-1b92ecac2777","Type":"ContainerStarted","Data":"cd16584ce87bf7e3a11abd9839b88969c56b57e5e4118d646432441bc2903fa5"} Mar 09 19:20:02 crc kubenswrapper[4750]: I0309 19:20:02.663805 4750 generic.go:334] "Generic (PLEG): container finished" podID="78e39153-00de-4622-a05a-1b92ecac2777" containerID="512ea9bad820cb6cd92ae8172e62dcae900c25448f4197e1479dd6ec609e1e7d" exitCode=0 Mar 09 19:20:02 crc kubenswrapper[4750]: I0309 19:20:02.663876 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551400-2jbpn" event={"ID":"78e39153-00de-4622-a05a-1b92ecac2777","Type":"ContainerDied","Data":"512ea9bad820cb6cd92ae8172e62dcae900c25448f4197e1479dd6ec609e1e7d"} Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.113775 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551400-2jbpn" Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.267281 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdvgb\" (UniqueName: \"kubernetes.io/projected/78e39153-00de-4622-a05a-1b92ecac2777-kube-api-access-sdvgb\") pod \"78e39153-00de-4622-a05a-1b92ecac2777\" (UID: \"78e39153-00de-4622-a05a-1b92ecac2777\") " Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.276039 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78e39153-00de-4622-a05a-1b92ecac2777-kube-api-access-sdvgb" (OuterVolumeSpecName: "kube-api-access-sdvgb") pod "78e39153-00de-4622-a05a-1b92ecac2777" (UID: "78e39153-00de-4622-a05a-1b92ecac2777"). InnerVolumeSpecName "kube-api-access-sdvgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.369486 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdvgb\" (UniqueName: \"kubernetes.io/projected/78e39153-00de-4622-a05a-1b92ecac2777-kube-api-access-sdvgb\") on node \"crc\" DevicePath \"\"" Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.372665 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:20:04 crc kubenswrapper[4750]: E0309 19:20:04.373011 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.690511 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551400-2jbpn" event={"ID":"78e39153-00de-4622-a05a-1b92ecac2777","Type":"ContainerDied","Data":"cd16584ce87bf7e3a11abd9839b88969c56b57e5e4118d646432441bc2903fa5"} Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.690579 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd16584ce87bf7e3a11abd9839b88969c56b57e5e4118d646432441bc2903fa5" Mar 09 19:20:04 crc kubenswrapper[4750]: I0309 19:20:04.690620 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551400-2jbpn" Mar 09 19:20:05 crc kubenswrapper[4750]: I0309 19:20:05.215861 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551394-svqlz"] Mar 09 19:20:05 crc kubenswrapper[4750]: I0309 19:20:05.231750 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551394-svqlz"] Mar 09 19:20:05 crc kubenswrapper[4750]: I0309 19:20:05.399225 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c683fa0a-42f4-4f36-9604-872c91065c80" path="/var/lib/kubelet/pods/c683fa0a-42f4-4f36-9604-872c91065c80/volumes" Mar 09 19:20:07 crc kubenswrapper[4750]: I0309 19:20:07.373804 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:20:07 crc kubenswrapper[4750]: E0309 19:20:07.374602 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:20:16 crc kubenswrapper[4750]: I0309 19:20:16.373883 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:20:16 crc kubenswrapper[4750]: E0309 19:20:16.375027 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:20:18 crc kubenswrapper[4750]: I0309 19:20:18.373915 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:20:18 crc kubenswrapper[4750]: E0309 19:20:18.374198 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:20:30 crc kubenswrapper[4750]: I0309 19:20:30.374180 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:20:30 crc kubenswrapper[4750]: E0309 19:20:30.375087 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:20:30 crc kubenswrapper[4750]: I0309 19:20:30.375280 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:20:30 crc kubenswrapper[4750]: E0309 19:20:30.375565 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:20:42 crc kubenswrapper[4750]: I0309 19:20:42.373949 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:20:42 crc kubenswrapper[4750]: E0309 19:20:42.375088 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:20:45 crc kubenswrapper[4750]: I0309 19:20:45.374561 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:20:45 crc kubenswrapper[4750]: E0309 19:20:45.375217 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:20:46 crc kubenswrapper[4750]: I0309 19:20:46.435311 4750 scope.go:117] "RemoveContainer" containerID="c7e67420a73aee4a5304d2e0ae787887c3143abce20047ec675cc0aa54f64c2b" Mar 09 19:20:55 crc kubenswrapper[4750]: I0309 19:20:55.373763 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:20:55 crc kubenswrapper[4750]: E0309 19:20:55.374386 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:21:00 crc kubenswrapper[4750]: I0309 19:21:00.374421 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:21:00 crc kubenswrapper[4750]: E0309 19:21:00.375217 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:21:07 crc kubenswrapper[4750]: I0309 19:21:07.374322 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:21:07 crc kubenswrapper[4750]: E0309 19:21:07.375513 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:21:14 crc kubenswrapper[4750]: I0309 19:21:14.373925 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:21:14 crc kubenswrapper[4750]: E0309 19:21:14.376015 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:21:15 crc kubenswrapper[4750]: I0309 19:21:15.909815 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jl9gw"] Mar 09 19:21:15 crc kubenswrapper[4750]: E0309 19:21:15.910581 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78e39153-00de-4622-a05a-1b92ecac2777" containerName="oc" Mar 09 19:21:15 crc kubenswrapper[4750]: I0309 19:21:15.910598 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="78e39153-00de-4622-a05a-1b92ecac2777" containerName="oc" Mar 09 19:21:15 crc kubenswrapper[4750]: I0309 19:21:15.910879 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="78e39153-00de-4622-a05a-1b92ecac2777" containerName="oc" Mar 09 19:21:15 crc kubenswrapper[4750]: I0309 19:21:15.912608 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:15 crc kubenswrapper[4750]: I0309 19:21:15.939219 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jl9gw"] Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.051822 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhkmw\" (UniqueName: \"kubernetes.io/projected/4416cc23-5fe9-4eb6-9630-6b27358d8552-kube-api-access-zhkmw\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.051884 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-catalog-content\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.052043 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-utilities\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.154158 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhkmw\" (UniqueName: \"kubernetes.io/projected/4416cc23-5fe9-4eb6-9630-6b27358d8552-kube-api-access-zhkmw\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.154425 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-catalog-content\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.154587 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-utilities\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.154929 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-catalog-content\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.155223 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-utilities\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.190678 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhkmw\" (UniqueName: \"kubernetes.io/projected/4416cc23-5fe9-4eb6-9630-6b27358d8552-kube-api-access-zhkmw\") pod \"redhat-operators-jl9gw\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.287083 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:16 crc kubenswrapper[4750]: I0309 19:21:16.802778 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jl9gw"] Mar 09 19:21:17 crc kubenswrapper[4750]: I0309 19:21:17.574724 4750 generic.go:334] "Generic (PLEG): container finished" podID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerID="2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f" exitCode=0 Mar 09 19:21:17 crc kubenswrapper[4750]: I0309 19:21:17.574768 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl9gw" event={"ID":"4416cc23-5fe9-4eb6-9630-6b27358d8552","Type":"ContainerDied","Data":"2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f"} Mar 09 19:21:17 crc kubenswrapper[4750]: I0309 19:21:17.574793 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl9gw" event={"ID":"4416cc23-5fe9-4eb6-9630-6b27358d8552","Type":"ContainerStarted","Data":"92b6749ea9ec9e6ba4e210acc4cab282ff14eb0a5d3f105d3086d020094228e1"} Mar 09 19:21:19 crc kubenswrapper[4750]: I0309 19:21:19.596926 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl9gw" event={"ID":"4416cc23-5fe9-4eb6-9630-6b27358d8552","Type":"ContainerStarted","Data":"1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c"} Mar 09 19:21:22 crc kubenswrapper[4750]: I0309 19:21:22.373819 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:21:22 crc kubenswrapper[4750]: E0309 19:21:22.375025 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:21:25 crc kubenswrapper[4750]: I0309 19:21:25.666956 4750 generic.go:334] "Generic (PLEG): container finished" podID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerID="1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c" exitCode=0 Mar 09 19:21:25 crc kubenswrapper[4750]: I0309 19:21:25.667065 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl9gw" event={"ID":"4416cc23-5fe9-4eb6-9630-6b27358d8552","Type":"ContainerDied","Data":"1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c"} Mar 09 19:21:26 crc kubenswrapper[4750]: I0309 19:21:26.373318 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:21:26 crc kubenswrapper[4750]: E0309 19:21:26.373785 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:21:26 crc kubenswrapper[4750]: I0309 19:21:26.681000 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl9gw" event={"ID":"4416cc23-5fe9-4eb6-9630-6b27358d8552","Type":"ContainerStarted","Data":"fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b"} Mar 09 19:21:26 crc kubenswrapper[4750]: I0309 19:21:26.703580 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jl9gw" podStartSLOduration=3.070118767 podStartE2EDuration="11.703559854s" podCreationTimestamp="2026-03-09 19:21:15 +0000 UTC" firstStartedPulling="2026-03-09 19:21:17.576081229 +0000 UTC m=+3358.918553627" lastFinishedPulling="2026-03-09 19:21:26.209522316 +0000 UTC m=+3367.551994714" observedRunningTime="2026-03-09 19:21:26.700393748 +0000 UTC m=+3368.042866156" watchObservedRunningTime="2026-03-09 19:21:26.703559854 +0000 UTC m=+3368.046032252" Mar 09 19:21:34 crc kubenswrapper[4750]: I0309 19:21:34.383383 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:21:34 crc kubenswrapper[4750]: E0309 19:21:34.384105 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:21:36 crc kubenswrapper[4750]: I0309 19:21:36.287360 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:36 crc kubenswrapper[4750]: I0309 19:21:36.287609 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:37 crc kubenswrapper[4750]: I0309 19:21:37.344301 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jl9gw" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="registry-server" probeResult="failure" output=< Mar 09 19:21:37 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:21:37 crc kubenswrapper[4750]: > Mar 09 19:21:38 crc kubenswrapper[4750]: I0309 19:21:38.373208 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:21:38 crc kubenswrapper[4750]: E0309 19:21:38.373698 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:21:45 crc kubenswrapper[4750]: I0309 19:21:45.374526 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:21:45 crc kubenswrapper[4750]: E0309 19:21:45.375575 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:21:46 crc kubenswrapper[4750]: I0309 19:21:46.348465 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:46 crc kubenswrapper[4750]: I0309 19:21:46.424028 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:47 crc kubenswrapper[4750]: I0309 19:21:47.123871 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jl9gw"] Mar 09 19:21:47 crc kubenswrapper[4750]: I0309 19:21:47.940432 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jl9gw" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="registry-server" containerID="cri-o://fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b" gracePeriod=2 Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.430562 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.623334 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-catalog-content\") pod \"4416cc23-5fe9-4eb6-9630-6b27358d8552\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.623831 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-utilities\") pod \"4416cc23-5fe9-4eb6-9630-6b27358d8552\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.624048 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhkmw\" (UniqueName: \"kubernetes.io/projected/4416cc23-5fe9-4eb6-9630-6b27358d8552-kube-api-access-zhkmw\") pod \"4416cc23-5fe9-4eb6-9630-6b27358d8552\" (UID: \"4416cc23-5fe9-4eb6-9630-6b27358d8552\") " Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.624919 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-utilities" (OuterVolumeSpecName: "utilities") pod "4416cc23-5fe9-4eb6-9630-6b27358d8552" (UID: "4416cc23-5fe9-4eb6-9630-6b27358d8552"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.631905 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4416cc23-5fe9-4eb6-9630-6b27358d8552-kube-api-access-zhkmw" (OuterVolumeSpecName: "kube-api-access-zhkmw") pod "4416cc23-5fe9-4eb6-9630-6b27358d8552" (UID: "4416cc23-5fe9-4eb6-9630-6b27358d8552"). InnerVolumeSpecName "kube-api-access-zhkmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.727198 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.727242 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhkmw\" (UniqueName: \"kubernetes.io/projected/4416cc23-5fe9-4eb6-9630-6b27358d8552-kube-api-access-zhkmw\") on node \"crc\" DevicePath \"\"" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.795079 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4416cc23-5fe9-4eb6-9630-6b27358d8552" (UID: "4416cc23-5fe9-4eb6-9630-6b27358d8552"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.829297 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4416cc23-5fe9-4eb6-9630-6b27358d8552-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.960187 4750 generic.go:334] "Generic (PLEG): container finished" podID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerID="fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b" exitCode=0 Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.960243 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl9gw" event={"ID":"4416cc23-5fe9-4eb6-9630-6b27358d8552","Type":"ContainerDied","Data":"fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b"} Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.960306 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl9gw" event={"ID":"4416cc23-5fe9-4eb6-9630-6b27358d8552","Type":"ContainerDied","Data":"92b6749ea9ec9e6ba4e210acc4cab282ff14eb0a5d3f105d3086d020094228e1"} Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.960311 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jl9gw" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.960335 4750 scope.go:117] "RemoveContainer" containerID="fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b" Mar 09 19:21:48 crc kubenswrapper[4750]: I0309 19:21:48.997910 4750 scope.go:117] "RemoveContainer" containerID="1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.001340 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jl9gw"] Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.012974 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jl9gw"] Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.029578 4750 scope.go:117] "RemoveContainer" containerID="2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.081027 4750 scope.go:117] "RemoveContainer" containerID="fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b" Mar 09 19:21:49 crc kubenswrapper[4750]: E0309 19:21:49.083659 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b\": container with ID starting with fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b not found: ID does not exist" containerID="fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.083703 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b"} err="failed to get container status \"fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b\": rpc error: code = NotFound desc = could not find container \"fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b\": container with ID starting with fdedc689c0d0ec473d6a97dd8d416baf63992225a879d20c289bdaf08c211c3b not found: ID does not exist" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.083737 4750 scope.go:117] "RemoveContainer" containerID="1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c" Mar 09 19:21:49 crc kubenswrapper[4750]: E0309 19:21:49.084445 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c\": container with ID starting with 1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c not found: ID does not exist" containerID="1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.084471 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c"} err="failed to get container status \"1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c\": rpc error: code = NotFound desc = could not find container \"1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c\": container with ID starting with 1573720d615bf54f38dc76239684fec9d39c57a6b7f5ed5ae90996522966da9c not found: ID does not exist" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.084489 4750 scope.go:117] "RemoveContainer" containerID="2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f" Mar 09 19:21:49 crc kubenswrapper[4750]: E0309 19:21:49.084797 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f\": container with ID starting with 2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f not found: ID does not exist" containerID="2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.084839 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f"} err="failed to get container status \"2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f\": rpc error: code = NotFound desc = could not find container \"2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f\": container with ID starting with 2bc09d02bc6ffc81016a8ba8f9432e3a58c11cc93be278ff4a2ba7bda917de4f not found: ID does not exist" Mar 09 19:21:49 crc kubenswrapper[4750]: I0309 19:21:49.382529 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" path="/var/lib/kubelet/pods/4416cc23-5fe9-4eb6-9630-6b27358d8552/volumes" Mar 09 19:21:51 crc kubenswrapper[4750]: I0309 19:21:51.374089 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:21:51 crc kubenswrapper[4750]: E0309 19:21:51.374552 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:21:58 crc kubenswrapper[4750]: I0309 19:21:58.374350 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:21:58 crc kubenswrapper[4750]: E0309 19:21:58.375425 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.169791 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551402-s4dt4"] Mar 09 19:22:00 crc kubenswrapper[4750]: E0309 19:22:00.170579 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="extract-content" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.170601 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="extract-content" Mar 09 19:22:00 crc kubenswrapper[4750]: E0309 19:22:00.170678 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="extract-utilities" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.170691 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="extract-utilities" Mar 09 19:22:00 crc kubenswrapper[4750]: E0309 19:22:00.170743 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="registry-server" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.170761 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="registry-server" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.171123 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4416cc23-5fe9-4eb6-9630-6b27358d8552" containerName="registry-server" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.172356 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551402-s4dt4" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.175186 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.175735 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.176091 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.185135 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551402-s4dt4"] Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.201145 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5jhk\" (UniqueName: \"kubernetes.io/projected/ee9777e2-f617-4509-9521-0c73b21737af-kube-api-access-j5jhk\") pod \"auto-csr-approver-29551402-s4dt4\" (UID: \"ee9777e2-f617-4509-9521-0c73b21737af\") " pod="openshift-infra/auto-csr-approver-29551402-s4dt4" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.302795 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5jhk\" (UniqueName: \"kubernetes.io/projected/ee9777e2-f617-4509-9521-0c73b21737af-kube-api-access-j5jhk\") pod \"auto-csr-approver-29551402-s4dt4\" (UID: \"ee9777e2-f617-4509-9521-0c73b21737af\") " pod="openshift-infra/auto-csr-approver-29551402-s4dt4" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.325541 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5jhk\" (UniqueName: \"kubernetes.io/projected/ee9777e2-f617-4509-9521-0c73b21737af-kube-api-access-j5jhk\") pod \"auto-csr-approver-29551402-s4dt4\" (UID: \"ee9777e2-f617-4509-9521-0c73b21737af\") " pod="openshift-infra/auto-csr-approver-29551402-s4dt4" Mar 09 19:22:00 crc kubenswrapper[4750]: I0309 19:22:00.498655 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551402-s4dt4" Mar 09 19:22:01 crc kubenswrapper[4750]: I0309 19:22:01.029103 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551402-s4dt4"] Mar 09 19:22:01 crc kubenswrapper[4750]: I0309 19:22:01.122886 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551402-s4dt4" event={"ID":"ee9777e2-f617-4509-9521-0c73b21737af","Type":"ContainerStarted","Data":"1910847b80ce9fafd8f887c03adfa0098d0f7bc5437e6b6adad02c52e87f135d"} Mar 09 19:22:03 crc kubenswrapper[4750]: I0309 19:22:03.146164 4750 generic.go:334] "Generic (PLEG): container finished" podID="ee9777e2-f617-4509-9521-0c73b21737af" containerID="d760fe8f5adf08f53a4dc21ee86ac830e427fc8d1f5d8e1c64d84681c2f1a893" exitCode=0 Mar 09 19:22:03 crc kubenswrapper[4750]: I0309 19:22:03.146700 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551402-s4dt4" event={"ID":"ee9777e2-f617-4509-9521-0c73b21737af","Type":"ContainerDied","Data":"d760fe8f5adf08f53a4dc21ee86ac830e427fc8d1f5d8e1c64d84681c2f1a893"} Mar 09 19:22:03 crc kubenswrapper[4750]: I0309 19:22:03.373904 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:22:03 crc kubenswrapper[4750]: E0309 19:22:03.374878 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:22:04 crc kubenswrapper[4750]: I0309 19:22:04.539448 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551402-s4dt4" Mar 09 19:22:04 crc kubenswrapper[4750]: I0309 19:22:04.741534 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5jhk\" (UniqueName: \"kubernetes.io/projected/ee9777e2-f617-4509-9521-0c73b21737af-kube-api-access-j5jhk\") pod \"ee9777e2-f617-4509-9521-0c73b21737af\" (UID: \"ee9777e2-f617-4509-9521-0c73b21737af\") " Mar 09 19:22:04 crc kubenswrapper[4750]: I0309 19:22:04.751020 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9777e2-f617-4509-9521-0c73b21737af-kube-api-access-j5jhk" (OuterVolumeSpecName: "kube-api-access-j5jhk") pod "ee9777e2-f617-4509-9521-0c73b21737af" (UID: "ee9777e2-f617-4509-9521-0c73b21737af"). InnerVolumeSpecName "kube-api-access-j5jhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:22:04 crc kubenswrapper[4750]: I0309 19:22:04.843958 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5jhk\" (UniqueName: \"kubernetes.io/projected/ee9777e2-f617-4509-9521-0c73b21737af-kube-api-access-j5jhk\") on node \"crc\" DevicePath \"\"" Mar 09 19:22:05 crc kubenswrapper[4750]: I0309 19:22:05.164534 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551402-s4dt4" event={"ID":"ee9777e2-f617-4509-9521-0c73b21737af","Type":"ContainerDied","Data":"1910847b80ce9fafd8f887c03adfa0098d0f7bc5437e6b6adad02c52e87f135d"} Mar 09 19:22:05 crc kubenswrapper[4750]: I0309 19:22:05.164799 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1910847b80ce9fafd8f887c03adfa0098d0f7bc5437e6b6adad02c52e87f135d" Mar 09 19:22:05 crc kubenswrapper[4750]: I0309 19:22:05.164736 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551402-s4dt4" Mar 09 19:22:05 crc kubenswrapper[4750]: I0309 19:22:05.629706 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551396-nqxcx"] Mar 09 19:22:05 crc kubenswrapper[4750]: I0309 19:22:05.641573 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551396-nqxcx"] Mar 09 19:22:07 crc kubenswrapper[4750]: I0309 19:22:07.385545 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff284095-d8e1-4ad3-9b21-a3ea610e448c" path="/var/lib/kubelet/pods/ff284095-d8e1-4ad3-9b21-a3ea610e448c/volumes" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.373838 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:22:13 crc kubenswrapper[4750]: E0309 19:22:13.374778 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.840007 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qtbjp"] Mar 09 19:22:13 crc kubenswrapper[4750]: E0309 19:22:13.840394 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9777e2-f617-4509-9521-0c73b21737af" containerName="oc" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.840405 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9777e2-f617-4509-9521-0c73b21737af" containerName="oc" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.840600 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9777e2-f617-4509-9521-0c73b21737af" containerName="oc" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.842092 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.857051 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qtbjp"] Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.964377 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-catalog-content\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.964587 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-utilities\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:13 crc kubenswrapper[4750]: I0309 19:22:13.964731 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9dzp\" (UniqueName: \"kubernetes.io/projected/7eceda2d-8784-47b1-b071-575acc550011-kube-api-access-l9dzp\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.066746 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-utilities\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.067045 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9dzp\" (UniqueName: \"kubernetes.io/projected/7eceda2d-8784-47b1-b071-575acc550011-kube-api-access-l9dzp\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.067187 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-catalog-content\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.067290 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-utilities\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.067566 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-catalog-content\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.087512 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9dzp\" (UniqueName: \"kubernetes.io/projected/7eceda2d-8784-47b1-b071-575acc550011-kube-api-access-l9dzp\") pod \"certified-operators-qtbjp\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.162055 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:14 crc kubenswrapper[4750]: I0309 19:22:14.799819 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qtbjp"] Mar 09 19:22:15 crc kubenswrapper[4750]: I0309 19:22:15.265138 4750 generic.go:334] "Generic (PLEG): container finished" podID="7eceda2d-8784-47b1-b071-575acc550011" containerID="24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741" exitCode=0 Mar 09 19:22:15 crc kubenswrapper[4750]: I0309 19:22:15.265188 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtbjp" event={"ID":"7eceda2d-8784-47b1-b071-575acc550011","Type":"ContainerDied","Data":"24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741"} Mar 09 19:22:15 crc kubenswrapper[4750]: I0309 19:22:15.265244 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtbjp" event={"ID":"7eceda2d-8784-47b1-b071-575acc550011","Type":"ContainerStarted","Data":"e10db7b0bf251173962265d7c1a1fbc31f7a0150d86ad7aa8e9f763cbd0f3b32"} Mar 09 19:22:16 crc kubenswrapper[4750]: I0309 19:22:16.276745 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtbjp" event={"ID":"7eceda2d-8784-47b1-b071-575acc550011","Type":"ContainerStarted","Data":"f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e"} Mar 09 19:22:17 crc kubenswrapper[4750]: I0309 19:22:17.292423 4750 generic.go:334] "Generic (PLEG): container finished" podID="7eceda2d-8784-47b1-b071-575acc550011" containerID="f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e" exitCode=0 Mar 09 19:22:17 crc kubenswrapper[4750]: I0309 19:22:17.292517 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtbjp" event={"ID":"7eceda2d-8784-47b1-b071-575acc550011","Type":"ContainerDied","Data":"f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e"} Mar 09 19:22:17 crc kubenswrapper[4750]: I0309 19:22:17.374007 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:22:17 crc kubenswrapper[4750]: E0309 19:22:17.374290 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:22:18 crc kubenswrapper[4750]: I0309 19:22:18.308623 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtbjp" event={"ID":"7eceda2d-8784-47b1-b071-575acc550011","Type":"ContainerStarted","Data":"5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9"} Mar 09 19:22:18 crc kubenswrapper[4750]: I0309 19:22:18.332523 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qtbjp" podStartSLOduration=2.854349624 podStartE2EDuration="5.332502417s" podCreationTimestamp="2026-03-09 19:22:13 +0000 UTC" firstStartedPulling="2026-03-09 19:22:15.266798709 +0000 UTC m=+3416.609271107" lastFinishedPulling="2026-03-09 19:22:17.744951502 +0000 UTC m=+3419.087423900" observedRunningTime="2026-03-09 19:22:18.325829447 +0000 UTC m=+3419.668301865" watchObservedRunningTime="2026-03-09 19:22:18.332502417 +0000 UTC m=+3419.674974825" Mar 09 19:22:24 crc kubenswrapper[4750]: I0309 19:22:24.817299 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:24 crc kubenswrapper[4750]: I0309 19:22:24.822242 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:24 crc kubenswrapper[4750]: I0309 19:22:24.825881 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:22:24 crc kubenswrapper[4750]: E0309 19:22:24.826457 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:22:24 crc kubenswrapper[4750]: I0309 19:22:24.897000 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:25 crc kubenswrapper[4750]: I0309 19:22:25.916820 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:25 crc kubenswrapper[4750]: I0309 19:22:25.979215 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qtbjp"] Mar 09 19:22:27 crc kubenswrapper[4750]: I0309 19:22:27.864775 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qtbjp" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="registry-server" containerID="cri-o://5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9" gracePeriod=2 Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.453955 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.627967 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-utilities\") pod \"7eceda2d-8784-47b1-b071-575acc550011\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.628045 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9dzp\" (UniqueName: \"kubernetes.io/projected/7eceda2d-8784-47b1-b071-575acc550011-kube-api-access-l9dzp\") pod \"7eceda2d-8784-47b1-b071-575acc550011\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.628139 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-catalog-content\") pod \"7eceda2d-8784-47b1-b071-575acc550011\" (UID: \"7eceda2d-8784-47b1-b071-575acc550011\") " Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.628954 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-utilities" (OuterVolumeSpecName: "utilities") pod "7eceda2d-8784-47b1-b071-575acc550011" (UID: "7eceda2d-8784-47b1-b071-575acc550011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.637146 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eceda2d-8784-47b1-b071-575acc550011-kube-api-access-l9dzp" (OuterVolumeSpecName: "kube-api-access-l9dzp") pod "7eceda2d-8784-47b1-b071-575acc550011" (UID: "7eceda2d-8784-47b1-b071-575acc550011"). InnerVolumeSpecName "kube-api-access-l9dzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.711964 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7eceda2d-8784-47b1-b071-575acc550011" (UID: "7eceda2d-8784-47b1-b071-575acc550011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.731368 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.731404 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9dzp\" (UniqueName: \"kubernetes.io/projected/7eceda2d-8784-47b1-b071-575acc550011-kube-api-access-l9dzp\") on node \"crc\" DevicePath \"\"" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.731420 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eceda2d-8784-47b1-b071-575acc550011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.876612 4750 generic.go:334] "Generic (PLEG): container finished" podID="7eceda2d-8784-47b1-b071-575acc550011" containerID="5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9" exitCode=0 Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.876666 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtbjp" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.877801 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtbjp" event={"ID":"7eceda2d-8784-47b1-b071-575acc550011","Type":"ContainerDied","Data":"5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9"} Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.877919 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtbjp" event={"ID":"7eceda2d-8784-47b1-b071-575acc550011","Type":"ContainerDied","Data":"e10db7b0bf251173962265d7c1a1fbc31f7a0150d86ad7aa8e9f763cbd0f3b32"} Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.878009 4750 scope.go:117] "RemoveContainer" containerID="5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.905878 4750 scope.go:117] "RemoveContainer" containerID="f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.937965 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qtbjp"] Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.946961 4750 scope.go:117] "RemoveContainer" containerID="24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.962431 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qtbjp"] Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.993837 4750 scope.go:117] "RemoveContainer" containerID="5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9" Mar 09 19:22:28 crc kubenswrapper[4750]: E0309 19:22:28.994462 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9\": container with ID starting with 5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9 not found: ID does not exist" containerID="5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.994511 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9"} err="failed to get container status \"5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9\": rpc error: code = NotFound desc = could not find container \"5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9\": container with ID starting with 5e5332533c4b523670a811c7a68d72444dcbfd53fb07ad43fe9b7fe870f0ddd9 not found: ID does not exist" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.994568 4750 scope.go:117] "RemoveContainer" containerID="f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e" Mar 09 19:22:28 crc kubenswrapper[4750]: E0309 19:22:28.995025 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e\": container with ID starting with f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e not found: ID does not exist" containerID="f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.995154 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e"} err="failed to get container status \"f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e\": rpc error: code = NotFound desc = could not find container \"f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e\": container with ID starting with f8e73b8acb5b26272f4e24043fb4915c14e3a12fa44640e32560f7e90d37756e not found: ID does not exist" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.995255 4750 scope.go:117] "RemoveContainer" containerID="24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741" Mar 09 19:22:28 crc kubenswrapper[4750]: E0309 19:22:28.995753 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741\": container with ID starting with 24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741 not found: ID does not exist" containerID="24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741" Mar 09 19:22:28 crc kubenswrapper[4750]: I0309 19:22:28.995865 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741"} err="failed to get container status \"24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741\": rpc error: code = NotFound desc = could not find container \"24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741\": container with ID starting with 24a423bc14a01450a3c590d1657b12927108ceb08522486dfd3f5f0f2f098741 not found: ID does not exist" Mar 09 19:22:29 crc kubenswrapper[4750]: I0309 19:22:29.410429 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eceda2d-8784-47b1-b071-575acc550011" path="/var/lib/kubelet/pods/7eceda2d-8784-47b1-b071-575acc550011/volumes" Mar 09 19:22:31 crc kubenswrapper[4750]: I0309 19:22:31.373829 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:22:31 crc kubenswrapper[4750]: E0309 19:22:31.374557 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:22:36 crc kubenswrapper[4750]: I0309 19:22:36.374662 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:22:36 crc kubenswrapper[4750]: E0309 19:22:36.375760 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:22:45 crc kubenswrapper[4750]: I0309 19:22:45.374678 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:22:45 crc kubenswrapper[4750]: E0309 19:22:45.375351 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:22:46 crc kubenswrapper[4750]: I0309 19:22:46.538023 4750 scope.go:117] "RemoveContainer" containerID="43eafd751ca9b9e68f2a74e9a5ec15856ae2d361796c923b4a721acfc9184fb7" Mar 09 19:22:48 crc kubenswrapper[4750]: I0309 19:22:48.373744 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:22:48 crc kubenswrapper[4750]: E0309 19:22:48.374579 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:23:00 crc kubenswrapper[4750]: I0309 19:23:00.375368 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:23:00 crc kubenswrapper[4750]: E0309 19:23:00.377394 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:23:02 crc kubenswrapper[4750]: I0309 19:23:02.373791 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:23:02 crc kubenswrapper[4750]: E0309 19:23:02.374330 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.631125 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zwr9d"] Mar 09 19:23:06 crc kubenswrapper[4750]: E0309 19:23:06.632475 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="extract-content" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.632497 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="extract-content" Mar 09 19:23:06 crc kubenswrapper[4750]: E0309 19:23:06.632546 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="extract-utilities" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.632561 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="extract-utilities" Mar 09 19:23:06 crc kubenswrapper[4750]: E0309 19:23:06.632579 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="registry-server" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.632592 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="registry-server" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.633014 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eceda2d-8784-47b1-b071-575acc550011" containerName="registry-server" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.635578 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.651469 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwr9d"] Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.767788 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2bh9\" (UniqueName: \"kubernetes.io/projected/b4e59834-bc15-4a90-8f75-ad858a090b9e-kube-api-access-z2bh9\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.768137 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-catalog-content\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.768227 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-utilities\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.870640 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2bh9\" (UniqueName: \"kubernetes.io/projected/b4e59834-bc15-4a90-8f75-ad858a090b9e-kube-api-access-z2bh9\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.870897 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-catalog-content\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.870980 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-utilities\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.871907 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-catalog-content\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.871976 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-utilities\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.903531 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2bh9\" (UniqueName: \"kubernetes.io/projected/b4e59834-bc15-4a90-8f75-ad858a090b9e-kube-api-access-z2bh9\") pod \"community-operators-zwr9d\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:06 crc kubenswrapper[4750]: I0309 19:23:06.967657 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:07 crc kubenswrapper[4750]: I0309 19:23:07.531982 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwr9d"] Mar 09 19:23:08 crc kubenswrapper[4750]: I0309 19:23:08.388113 4750 generic.go:334] "Generic (PLEG): container finished" podID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerID="74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e" exitCode=0 Mar 09 19:23:08 crc kubenswrapper[4750]: I0309 19:23:08.388200 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwr9d" event={"ID":"b4e59834-bc15-4a90-8f75-ad858a090b9e","Type":"ContainerDied","Data":"74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e"} Mar 09 19:23:08 crc kubenswrapper[4750]: I0309 19:23:08.388510 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwr9d" event={"ID":"b4e59834-bc15-4a90-8f75-ad858a090b9e","Type":"ContainerStarted","Data":"a2204554f6db0b648f96376ce187431c601e81ff433d3bdedc41aa7fe11c1971"} Mar 09 19:23:10 crc kubenswrapper[4750]: I0309 19:23:10.416560 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwr9d" event={"ID":"b4e59834-bc15-4a90-8f75-ad858a090b9e","Type":"ContainerStarted","Data":"9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642"} Mar 09 19:23:11 crc kubenswrapper[4750]: I0309 19:23:11.428996 4750 generic.go:334] "Generic (PLEG): container finished" podID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerID="9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642" exitCode=0 Mar 09 19:23:11 crc kubenswrapper[4750]: I0309 19:23:11.429101 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwr9d" event={"ID":"b4e59834-bc15-4a90-8f75-ad858a090b9e","Type":"ContainerDied","Data":"9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642"} Mar 09 19:23:12 crc kubenswrapper[4750]: I0309 19:23:12.443114 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwr9d" event={"ID":"b4e59834-bc15-4a90-8f75-ad858a090b9e","Type":"ContainerStarted","Data":"ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47"} Mar 09 19:23:12 crc kubenswrapper[4750]: I0309 19:23:12.467876 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zwr9d" podStartSLOduration=3.041055055 podStartE2EDuration="6.467855277s" podCreationTimestamp="2026-03-09 19:23:06 +0000 UTC" firstStartedPulling="2026-03-09 19:23:08.391644277 +0000 UTC m=+3469.734116685" lastFinishedPulling="2026-03-09 19:23:11.818444479 +0000 UTC m=+3473.160916907" observedRunningTime="2026-03-09 19:23:12.462531343 +0000 UTC m=+3473.805003781" watchObservedRunningTime="2026-03-09 19:23:12.467855277 +0000 UTC m=+3473.810327665" Mar 09 19:23:14 crc kubenswrapper[4750]: I0309 19:23:14.374840 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:23:14 crc kubenswrapper[4750]: E0309 19:23:14.375616 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:23:16 crc kubenswrapper[4750]: I0309 19:23:16.968799 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:16 crc kubenswrapper[4750]: I0309 19:23:16.968891 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:17 crc kubenswrapper[4750]: I0309 19:23:17.044895 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:17 crc kubenswrapper[4750]: I0309 19:23:17.374753 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:23:17 crc kubenswrapper[4750]: E0309 19:23:17.375134 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:23:17 crc kubenswrapper[4750]: I0309 19:23:17.581100 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:18 crc kubenswrapper[4750]: I0309 19:23:18.815870 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwr9d"] Mar 09 19:23:19 crc kubenswrapper[4750]: I0309 19:23:19.514418 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zwr9d" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="registry-server" containerID="cri-o://ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47" gracePeriod=2 Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.026975 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.098131 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-utilities\") pod \"b4e59834-bc15-4a90-8f75-ad858a090b9e\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.098252 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-catalog-content\") pod \"b4e59834-bc15-4a90-8f75-ad858a090b9e\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.098394 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2bh9\" (UniqueName: \"kubernetes.io/projected/b4e59834-bc15-4a90-8f75-ad858a090b9e-kube-api-access-z2bh9\") pod \"b4e59834-bc15-4a90-8f75-ad858a090b9e\" (UID: \"b4e59834-bc15-4a90-8f75-ad858a090b9e\") " Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.099651 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-utilities" (OuterVolumeSpecName: "utilities") pod "b4e59834-bc15-4a90-8f75-ad858a090b9e" (UID: "b4e59834-bc15-4a90-8f75-ad858a090b9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.106967 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e59834-bc15-4a90-8f75-ad858a090b9e-kube-api-access-z2bh9" (OuterVolumeSpecName: "kube-api-access-z2bh9") pod "b4e59834-bc15-4a90-8f75-ad858a090b9e" (UID: "b4e59834-bc15-4a90-8f75-ad858a090b9e"). InnerVolumeSpecName "kube-api-access-z2bh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.201173 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2bh9\" (UniqueName: \"kubernetes.io/projected/b4e59834-bc15-4a90-8f75-ad858a090b9e-kube-api-access-z2bh9\") on node \"crc\" DevicePath \"\"" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.201432 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.242518 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4e59834-bc15-4a90-8f75-ad858a090b9e" (UID: "b4e59834-bc15-4a90-8f75-ad858a090b9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.304092 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e59834-bc15-4a90-8f75-ad858a090b9e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.527914 4750 generic.go:334] "Generic (PLEG): container finished" podID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerID="ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47" exitCode=0 Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.527953 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwr9d" event={"ID":"b4e59834-bc15-4a90-8f75-ad858a090b9e","Type":"ContainerDied","Data":"ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47"} Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.527981 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwr9d" event={"ID":"b4e59834-bc15-4a90-8f75-ad858a090b9e","Type":"ContainerDied","Data":"a2204554f6db0b648f96376ce187431c601e81ff433d3bdedc41aa7fe11c1971"} Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.527999 4750 scope.go:117] "RemoveContainer" containerID="ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.528040 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwr9d" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.567194 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwr9d"] Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.574793 4750 scope.go:117] "RemoveContainer" containerID="9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.587183 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zwr9d"] Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.597090 4750 scope.go:117] "RemoveContainer" containerID="74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.651556 4750 scope.go:117] "RemoveContainer" containerID="ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47" Mar 09 19:23:20 crc kubenswrapper[4750]: E0309 19:23:20.652233 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47\": container with ID starting with ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47 not found: ID does not exist" containerID="ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.652276 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47"} err="failed to get container status \"ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47\": rpc error: code = NotFound desc = could not find container \"ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47\": container with ID starting with ec09fa8333b0134c82bf141c99fe631e949f8498b7e3076e3b3a526eca324d47 not found: ID does not exist" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.652307 4750 scope.go:117] "RemoveContainer" containerID="9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642" Mar 09 19:23:20 crc kubenswrapper[4750]: E0309 19:23:20.652749 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642\": container with ID starting with 9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642 not found: ID does not exist" containerID="9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.652884 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642"} err="failed to get container status \"9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642\": rpc error: code = NotFound desc = could not find container \"9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642\": container with ID starting with 9e0e49e4759791a106822b429772e9ea771b1dd09e47a691aaf665e6d317f642 not found: ID does not exist" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.652985 4750 scope.go:117] "RemoveContainer" containerID="74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e" Mar 09 19:23:20 crc kubenswrapper[4750]: E0309 19:23:20.653372 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e\": container with ID starting with 74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e not found: ID does not exist" containerID="74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e" Mar 09 19:23:20 crc kubenswrapper[4750]: I0309 19:23:20.653431 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e"} err="failed to get container status \"74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e\": rpc error: code = NotFound desc = could not find container \"74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e\": container with ID starting with 74190169a124f33515250b389dd4705821ddd8c9a74289828fd3c3a8ca288c8e not found: ID does not exist" Mar 09 19:23:21 crc kubenswrapper[4750]: I0309 19:23:21.391926 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" path="/var/lib/kubelet/pods/b4e59834-bc15-4a90-8f75-ad858a090b9e/volumes" Mar 09 19:23:25 crc kubenswrapper[4750]: I0309 19:23:25.375672 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:23:25 crc kubenswrapper[4750]: E0309 19:23:25.376766 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:23:31 crc kubenswrapper[4750]: I0309 19:23:31.374327 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:23:31 crc kubenswrapper[4750]: I0309 19:23:31.682176 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"09feee34ec58900b16a230485b577a5b4f6613c80d8a21dc529b58c82458c699"} Mar 09 19:23:36 crc kubenswrapper[4750]: I0309 19:23:36.374354 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:23:36 crc kubenswrapper[4750]: E0309 19:23:36.375377 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:23:48 crc kubenswrapper[4750]: I0309 19:23:48.374746 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:23:48 crc kubenswrapper[4750]: E0309 19:23:48.375954 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.157386 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551404-7pwhk"] Mar 09 19:24:00 crc kubenswrapper[4750]: E0309 19:24:00.158662 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="extract-utilities" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.158697 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="extract-utilities" Mar 09 19:24:00 crc kubenswrapper[4750]: E0309 19:24:00.158769 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="extract-content" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.158783 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="extract-content" Mar 09 19:24:00 crc kubenswrapper[4750]: E0309 19:24:00.158811 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="registry-server" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.158827 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="registry-server" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.159202 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e59834-bc15-4a90-8f75-ad858a090b9e" containerName="registry-server" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.160411 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551404-7pwhk" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.164571 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.164605 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.164733 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.185829 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551404-7pwhk"] Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.270809 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6djb\" (UniqueName: \"kubernetes.io/projected/1523b29c-ce7d-45fc-88fd-4192c9dc2e28-kube-api-access-j6djb\") pod \"auto-csr-approver-29551404-7pwhk\" (UID: \"1523b29c-ce7d-45fc-88fd-4192c9dc2e28\") " pod="openshift-infra/auto-csr-approver-29551404-7pwhk" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.373187 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6djb\" (UniqueName: \"kubernetes.io/projected/1523b29c-ce7d-45fc-88fd-4192c9dc2e28-kube-api-access-j6djb\") pod \"auto-csr-approver-29551404-7pwhk\" (UID: \"1523b29c-ce7d-45fc-88fd-4192c9dc2e28\") " pod="openshift-infra/auto-csr-approver-29551404-7pwhk" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.404064 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6djb\" (UniqueName: \"kubernetes.io/projected/1523b29c-ce7d-45fc-88fd-4192c9dc2e28-kube-api-access-j6djb\") pod \"auto-csr-approver-29551404-7pwhk\" (UID: \"1523b29c-ce7d-45fc-88fd-4192c9dc2e28\") " pod="openshift-infra/auto-csr-approver-29551404-7pwhk" Mar 09 19:24:00 crc kubenswrapper[4750]: I0309 19:24:00.485229 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551404-7pwhk" Mar 09 19:24:01 crc kubenswrapper[4750]: W0309 19:24:01.042218 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1523b29c_ce7d_45fc_88fd_4192c9dc2e28.slice/crio-7abeaa1245e8e3a046d252117629bd74a5b5ab24434e18113f0cb2378b565096 WatchSource:0}: Error finding container 7abeaa1245e8e3a046d252117629bd74a5b5ab24434e18113f0cb2378b565096: Status 404 returned error can't find the container with id 7abeaa1245e8e3a046d252117629bd74a5b5ab24434e18113f0cb2378b565096 Mar 09 19:24:01 crc kubenswrapper[4750]: I0309 19:24:01.046377 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551404-7pwhk"] Mar 09 19:24:01 crc kubenswrapper[4750]: I0309 19:24:01.095201 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551404-7pwhk" event={"ID":"1523b29c-ce7d-45fc-88fd-4192c9dc2e28","Type":"ContainerStarted","Data":"7abeaa1245e8e3a046d252117629bd74a5b5ab24434e18113f0cb2378b565096"} Mar 09 19:24:01 crc kubenswrapper[4750]: I0309 19:24:01.375907 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:24:01 crc kubenswrapper[4750]: E0309 19:24:01.376514 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:24:03 crc kubenswrapper[4750]: I0309 19:24:03.119666 4750 generic.go:334] "Generic (PLEG): container finished" podID="1523b29c-ce7d-45fc-88fd-4192c9dc2e28" containerID="699a78d92c455e1d9a61a1e084d6c2515943ed9179e6f8403fa5c782d36433db" exitCode=0 Mar 09 19:24:03 crc kubenswrapper[4750]: I0309 19:24:03.119758 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551404-7pwhk" event={"ID":"1523b29c-ce7d-45fc-88fd-4192c9dc2e28","Type":"ContainerDied","Data":"699a78d92c455e1d9a61a1e084d6c2515943ed9179e6f8403fa5c782d36433db"} Mar 09 19:24:04 crc kubenswrapper[4750]: I0309 19:24:04.520943 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551404-7pwhk" Mar 09 19:24:04 crc kubenswrapper[4750]: I0309 19:24:04.598383 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6djb\" (UniqueName: \"kubernetes.io/projected/1523b29c-ce7d-45fc-88fd-4192c9dc2e28-kube-api-access-j6djb\") pod \"1523b29c-ce7d-45fc-88fd-4192c9dc2e28\" (UID: \"1523b29c-ce7d-45fc-88fd-4192c9dc2e28\") " Mar 09 19:24:04 crc kubenswrapper[4750]: I0309 19:24:04.607110 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1523b29c-ce7d-45fc-88fd-4192c9dc2e28-kube-api-access-j6djb" (OuterVolumeSpecName: "kube-api-access-j6djb") pod "1523b29c-ce7d-45fc-88fd-4192c9dc2e28" (UID: "1523b29c-ce7d-45fc-88fd-4192c9dc2e28"). InnerVolumeSpecName "kube-api-access-j6djb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:24:04 crc kubenswrapper[4750]: I0309 19:24:04.700968 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6djb\" (UniqueName: \"kubernetes.io/projected/1523b29c-ce7d-45fc-88fd-4192c9dc2e28-kube-api-access-j6djb\") on node \"crc\" DevicePath \"\"" Mar 09 19:24:05 crc kubenswrapper[4750]: I0309 19:24:05.141788 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551404-7pwhk" event={"ID":"1523b29c-ce7d-45fc-88fd-4192c9dc2e28","Type":"ContainerDied","Data":"7abeaa1245e8e3a046d252117629bd74a5b5ab24434e18113f0cb2378b565096"} Mar 09 19:24:05 crc kubenswrapper[4750]: I0309 19:24:05.142162 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7abeaa1245e8e3a046d252117629bd74a5b5ab24434e18113f0cb2378b565096" Mar 09 19:24:05 crc kubenswrapper[4750]: I0309 19:24:05.141902 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551404-7pwhk" Mar 09 19:24:05 crc kubenswrapper[4750]: I0309 19:24:05.619935 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551398-znfd4"] Mar 09 19:24:05 crc kubenswrapper[4750]: I0309 19:24:05.629344 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551398-znfd4"] Mar 09 19:24:07 crc kubenswrapper[4750]: I0309 19:24:07.387364 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4afddec7-7283-410f-8707-2f4112be0f91" path="/var/lib/kubelet/pods/4afddec7-7283-410f-8707-2f4112be0f91/volumes" Mar 09 19:24:12 crc kubenswrapper[4750]: I0309 19:24:12.373678 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:24:12 crc kubenswrapper[4750]: E0309 19:24:12.374483 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:24:26 crc kubenswrapper[4750]: I0309 19:24:26.374349 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:24:26 crc kubenswrapper[4750]: E0309 19:24:26.375611 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:24:40 crc kubenswrapper[4750]: I0309 19:24:40.374079 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:24:40 crc kubenswrapper[4750]: E0309 19:24:40.375187 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:24:46 crc kubenswrapper[4750]: I0309 19:24:46.747568 4750 scope.go:117] "RemoveContainer" containerID="de48542d81c0694d3c1764638d52e91e4bc6e50c4ae371f202fc275da0becacb" Mar 09 19:24:51 crc kubenswrapper[4750]: I0309 19:24:51.375570 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:24:51 crc kubenswrapper[4750]: E0309 19:24:51.376823 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:25:02 crc kubenswrapper[4750]: I0309 19:25:02.373513 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:25:03 crc kubenswrapper[4750]: I0309 19:25:03.876886 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44"} Mar 09 19:25:05 crc kubenswrapper[4750]: I0309 19:25:05.190276 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:25:05 crc kubenswrapper[4750]: I0309 19:25:05.904930 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" exitCode=1 Mar 09 19:25:05 crc kubenswrapper[4750]: I0309 19:25:05.905007 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44"} Mar 09 19:25:05 crc kubenswrapper[4750]: I0309 19:25:05.905303 4750 scope.go:117] "RemoveContainer" containerID="9399c29b984e0749005a13098f83fe817a3ac655fb93e40ea016c5a312908df5" Mar 09 19:25:05 crc kubenswrapper[4750]: I0309 19:25:05.906533 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:25:05 crc kubenswrapper[4750]: E0309 19:25:05.907308 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:25:08 crc kubenswrapper[4750]: I0309 19:25:08.189416 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:25:08 crc kubenswrapper[4750]: I0309 19:25:08.190748 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:25:08 crc kubenswrapper[4750]: E0309 19:25:08.191220 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:25:10 crc kubenswrapper[4750]: I0309 19:25:10.190197 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:25:10 crc kubenswrapper[4750]: I0309 19:25:10.191370 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:25:10 crc kubenswrapper[4750]: E0309 19:25:10.191741 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:25:23 crc kubenswrapper[4750]: I0309 19:25:23.373619 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:25:23 crc kubenswrapper[4750]: E0309 19:25:23.374608 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.482776 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gmhmq"] Mar 09 19:25:30 crc kubenswrapper[4750]: E0309 19:25:30.483994 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1523b29c-ce7d-45fc-88fd-4192c9dc2e28" containerName="oc" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.484015 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1523b29c-ce7d-45fc-88fd-4192c9dc2e28" containerName="oc" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.484407 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1523b29c-ce7d-45fc-88fd-4192c9dc2e28" containerName="oc" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.486984 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.503711 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gmhmq"] Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.627337 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-catalog-content\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.627912 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-utilities\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.628341 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbb9h\" (UniqueName: \"kubernetes.io/projected/97906a08-3691-4374-b674-ac440f9359a6-kube-api-access-lbb9h\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.730671 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbb9h\" (UniqueName: \"kubernetes.io/projected/97906a08-3691-4374-b674-ac440f9359a6-kube-api-access-lbb9h\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.731093 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-catalog-content\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.731128 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-utilities\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.731578 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-catalog-content\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.731669 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-utilities\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.759149 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbb9h\" (UniqueName: \"kubernetes.io/projected/97906a08-3691-4374-b674-ac440f9359a6-kube-api-access-lbb9h\") pod \"redhat-marketplace-gmhmq\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:30 crc kubenswrapper[4750]: I0309 19:25:30.835170 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:31 crc kubenswrapper[4750]: W0309 19:25:31.316825 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97906a08_3691_4374_b674_ac440f9359a6.slice/crio-8daea03452b012125e2ac27c2ae9389076a23642ae8a6c33df0244f9a88aeeb6 WatchSource:0}: Error finding container 8daea03452b012125e2ac27c2ae9389076a23642ae8a6c33df0244f9a88aeeb6: Status 404 returned error can't find the container with id 8daea03452b012125e2ac27c2ae9389076a23642ae8a6c33df0244f9a88aeeb6 Mar 09 19:25:31 crc kubenswrapper[4750]: I0309 19:25:31.323435 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gmhmq"] Mar 09 19:25:32 crc kubenswrapper[4750]: I0309 19:25:32.252701 4750 generic.go:334] "Generic (PLEG): container finished" podID="97906a08-3691-4374-b674-ac440f9359a6" containerID="60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f" exitCode=0 Mar 09 19:25:32 crc kubenswrapper[4750]: I0309 19:25:32.252893 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gmhmq" event={"ID":"97906a08-3691-4374-b674-ac440f9359a6","Type":"ContainerDied","Data":"60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f"} Mar 09 19:25:32 crc kubenswrapper[4750]: I0309 19:25:32.253202 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gmhmq" event={"ID":"97906a08-3691-4374-b674-ac440f9359a6","Type":"ContainerStarted","Data":"8daea03452b012125e2ac27c2ae9389076a23642ae8a6c33df0244f9a88aeeb6"} Mar 09 19:25:32 crc kubenswrapper[4750]: I0309 19:25:32.259164 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:25:34 crc kubenswrapper[4750]: I0309 19:25:34.374896 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:25:34 crc kubenswrapper[4750]: E0309 19:25:34.375818 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:25:35 crc kubenswrapper[4750]: I0309 19:25:35.292892 4750 generic.go:334] "Generic (PLEG): container finished" podID="97906a08-3691-4374-b674-ac440f9359a6" containerID="70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6" exitCode=0 Mar 09 19:25:35 crc kubenswrapper[4750]: I0309 19:25:35.293261 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gmhmq" event={"ID":"97906a08-3691-4374-b674-ac440f9359a6","Type":"ContainerDied","Data":"70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6"} Mar 09 19:25:36 crc kubenswrapper[4750]: I0309 19:25:36.305285 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gmhmq" event={"ID":"97906a08-3691-4374-b674-ac440f9359a6","Type":"ContainerStarted","Data":"1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839"} Mar 09 19:25:36 crc kubenswrapper[4750]: I0309 19:25:36.332035 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gmhmq" podStartSLOduration=2.894986105 podStartE2EDuration="6.332018309s" podCreationTimestamp="2026-03-09 19:25:30 +0000 UTC" firstStartedPulling="2026-03-09 19:25:32.258902914 +0000 UTC m=+3613.601375332" lastFinishedPulling="2026-03-09 19:25:35.695935138 +0000 UTC m=+3617.038407536" observedRunningTime="2026-03-09 19:25:36.324539876 +0000 UTC m=+3617.667012274" watchObservedRunningTime="2026-03-09 19:25:36.332018309 +0000 UTC m=+3617.674490707" Mar 09 19:25:40 crc kubenswrapper[4750]: I0309 19:25:40.835358 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:40 crc kubenswrapper[4750]: I0309 19:25:40.836066 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:40 crc kubenswrapper[4750]: I0309 19:25:40.927521 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:41 crc kubenswrapper[4750]: I0309 19:25:41.459725 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:41 crc kubenswrapper[4750]: I0309 19:25:41.524087 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gmhmq"] Mar 09 19:25:43 crc kubenswrapper[4750]: I0309 19:25:43.399371 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gmhmq" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="registry-server" containerID="cri-o://1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839" gracePeriod=2 Mar 09 19:25:43 crc kubenswrapper[4750]: I0309 19:25:43.992852 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.048715 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-utilities\") pod \"97906a08-3691-4374-b674-ac440f9359a6\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.048785 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-catalog-content\") pod \"97906a08-3691-4374-b674-ac440f9359a6\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.048996 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbb9h\" (UniqueName: \"kubernetes.io/projected/97906a08-3691-4374-b674-ac440f9359a6-kube-api-access-lbb9h\") pod \"97906a08-3691-4374-b674-ac440f9359a6\" (UID: \"97906a08-3691-4374-b674-ac440f9359a6\") " Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.049616 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-utilities" (OuterVolumeSpecName: "utilities") pod "97906a08-3691-4374-b674-ac440f9359a6" (UID: "97906a08-3691-4374-b674-ac440f9359a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.052319 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.058447 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97906a08-3691-4374-b674-ac440f9359a6-kube-api-access-lbb9h" (OuterVolumeSpecName: "kube-api-access-lbb9h") pod "97906a08-3691-4374-b674-ac440f9359a6" (UID: "97906a08-3691-4374-b674-ac440f9359a6"). InnerVolumeSpecName "kube-api-access-lbb9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.086515 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97906a08-3691-4374-b674-ac440f9359a6" (UID: "97906a08-3691-4374-b674-ac440f9359a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.153926 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97906a08-3691-4374-b674-ac440f9359a6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.153958 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbb9h\" (UniqueName: \"kubernetes.io/projected/97906a08-3691-4374-b674-ac440f9359a6-kube-api-access-lbb9h\") on node \"crc\" DevicePath \"\"" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.412954 4750 generic.go:334] "Generic (PLEG): container finished" podID="97906a08-3691-4374-b674-ac440f9359a6" containerID="1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839" exitCode=0 Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.413020 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gmhmq" event={"ID":"97906a08-3691-4374-b674-ac440f9359a6","Type":"ContainerDied","Data":"1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839"} Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.413521 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gmhmq" event={"ID":"97906a08-3691-4374-b674-ac440f9359a6","Type":"ContainerDied","Data":"8daea03452b012125e2ac27c2ae9389076a23642ae8a6c33df0244f9a88aeeb6"} Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.413549 4750 scope.go:117] "RemoveContainer" containerID="1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.413170 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gmhmq" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.436807 4750 scope.go:117] "RemoveContainer" containerID="70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.454962 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gmhmq"] Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.465055 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gmhmq"] Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.480451 4750 scope.go:117] "RemoveContainer" containerID="60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.509557 4750 scope.go:117] "RemoveContainer" containerID="1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839" Mar 09 19:25:44 crc kubenswrapper[4750]: E0309 19:25:44.510021 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839\": container with ID starting with 1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839 not found: ID does not exist" containerID="1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.510055 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839"} err="failed to get container status \"1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839\": rpc error: code = NotFound desc = could not find container \"1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839\": container with ID starting with 1367f94f4f1af1c290ac8af1041d61aa5836608d5db18792622645f44bce3839 not found: ID does not exist" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.510079 4750 scope.go:117] "RemoveContainer" containerID="70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6" Mar 09 19:25:44 crc kubenswrapper[4750]: E0309 19:25:44.510519 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6\": container with ID starting with 70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6 not found: ID does not exist" containerID="70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.510553 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6"} err="failed to get container status \"70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6\": rpc error: code = NotFound desc = could not find container \"70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6\": container with ID starting with 70b518ee5be82b35ec248186d10fb086e23d3adacb6d5113e774ec1df957a2a6 not found: ID does not exist" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.510571 4750 scope.go:117] "RemoveContainer" containerID="60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f" Mar 09 19:25:44 crc kubenswrapper[4750]: E0309 19:25:44.510977 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f\": container with ID starting with 60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f not found: ID does not exist" containerID="60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f" Mar 09 19:25:44 crc kubenswrapper[4750]: I0309 19:25:44.511002 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f"} err="failed to get container status \"60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f\": rpc error: code = NotFound desc = could not find container \"60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f\": container with ID starting with 60b2700859d3ab9a7190cd8b16e270ed9b98474706dc414e040a33e50a0ccc8f not found: ID does not exist" Mar 09 19:25:45 crc kubenswrapper[4750]: I0309 19:25:45.397359 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97906a08-3691-4374-b674-ac440f9359a6" path="/var/lib/kubelet/pods/97906a08-3691-4374-b674-ac440f9359a6/volumes" Mar 09 19:25:49 crc kubenswrapper[4750]: I0309 19:25:49.379529 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:25:49 crc kubenswrapper[4750]: E0309 19:25:49.380278 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:25:51 crc kubenswrapper[4750]: I0309 19:25:51.744158 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:25:51 crc kubenswrapper[4750]: I0309 19:25:51.744514 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.163429 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551406-wxvp8"] Mar 09 19:26:00 crc kubenswrapper[4750]: E0309 19:26:00.164525 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="extract-content" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.164544 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="extract-content" Mar 09 19:26:00 crc kubenswrapper[4750]: E0309 19:26:00.164575 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="extract-utilities" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.164584 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="extract-utilities" Mar 09 19:26:00 crc kubenswrapper[4750]: E0309 19:26:00.164612 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="registry-server" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.164621 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="registry-server" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.164913 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="97906a08-3691-4374-b674-ac440f9359a6" containerName="registry-server" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.166022 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551406-wxvp8" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.171138 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.171315 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.171426 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.181534 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551406-wxvp8"] Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.224514 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5h6n\" (UniqueName: \"kubernetes.io/projected/e8e33970-767e-40c7-a5ba-11ad5473b535-kube-api-access-l5h6n\") pod \"auto-csr-approver-29551406-wxvp8\" (UID: \"e8e33970-767e-40c7-a5ba-11ad5473b535\") " pod="openshift-infra/auto-csr-approver-29551406-wxvp8" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.327486 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5h6n\" (UniqueName: \"kubernetes.io/projected/e8e33970-767e-40c7-a5ba-11ad5473b535-kube-api-access-l5h6n\") pod \"auto-csr-approver-29551406-wxvp8\" (UID: \"e8e33970-767e-40c7-a5ba-11ad5473b535\") " pod="openshift-infra/auto-csr-approver-29551406-wxvp8" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.367516 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5h6n\" (UniqueName: \"kubernetes.io/projected/e8e33970-767e-40c7-a5ba-11ad5473b535-kube-api-access-l5h6n\") pod \"auto-csr-approver-29551406-wxvp8\" (UID: \"e8e33970-767e-40c7-a5ba-11ad5473b535\") " pod="openshift-infra/auto-csr-approver-29551406-wxvp8" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.373342 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:26:00 crc kubenswrapper[4750]: E0309 19:26:00.373901 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:26:00 crc kubenswrapper[4750]: I0309 19:26:00.491094 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551406-wxvp8" Mar 09 19:26:01 crc kubenswrapper[4750]: I0309 19:26:01.098885 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551406-wxvp8"] Mar 09 19:26:01 crc kubenswrapper[4750]: I0309 19:26:01.616546 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551406-wxvp8" event={"ID":"e8e33970-767e-40c7-a5ba-11ad5473b535","Type":"ContainerStarted","Data":"d6f1448dedfa4de54eabd614cea20bd7a12db30bb599b3142b4f3befaa84637c"} Mar 09 19:26:03 crc kubenswrapper[4750]: I0309 19:26:03.638170 4750 generic.go:334] "Generic (PLEG): container finished" podID="e8e33970-767e-40c7-a5ba-11ad5473b535" containerID="af19d8bc0656e16d31076a3067cb291b01619330b0337ea9656f1b94fa51124e" exitCode=0 Mar 09 19:26:03 crc kubenswrapper[4750]: I0309 19:26:03.638277 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551406-wxvp8" event={"ID":"e8e33970-767e-40c7-a5ba-11ad5473b535","Type":"ContainerDied","Data":"af19d8bc0656e16d31076a3067cb291b01619330b0337ea9656f1b94fa51124e"} Mar 09 19:26:05 crc kubenswrapper[4750]: I0309 19:26:05.058701 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551406-wxvp8" Mar 09 19:26:05 crc kubenswrapper[4750]: I0309 19:26:05.136358 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5h6n\" (UniqueName: \"kubernetes.io/projected/e8e33970-767e-40c7-a5ba-11ad5473b535-kube-api-access-l5h6n\") pod \"e8e33970-767e-40c7-a5ba-11ad5473b535\" (UID: \"e8e33970-767e-40c7-a5ba-11ad5473b535\") " Mar 09 19:26:05 crc kubenswrapper[4750]: I0309 19:26:05.143661 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8e33970-767e-40c7-a5ba-11ad5473b535-kube-api-access-l5h6n" (OuterVolumeSpecName: "kube-api-access-l5h6n") pod "e8e33970-767e-40c7-a5ba-11ad5473b535" (UID: "e8e33970-767e-40c7-a5ba-11ad5473b535"). InnerVolumeSpecName "kube-api-access-l5h6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:26:05 crc kubenswrapper[4750]: I0309 19:26:05.238459 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5h6n\" (UniqueName: \"kubernetes.io/projected/e8e33970-767e-40c7-a5ba-11ad5473b535-kube-api-access-l5h6n\") on node \"crc\" DevicePath \"\"" Mar 09 19:26:05 crc kubenswrapper[4750]: I0309 19:26:05.662358 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551406-wxvp8" event={"ID":"e8e33970-767e-40c7-a5ba-11ad5473b535","Type":"ContainerDied","Data":"d6f1448dedfa4de54eabd614cea20bd7a12db30bb599b3142b4f3befaa84637c"} Mar 09 19:26:05 crc kubenswrapper[4750]: I0309 19:26:05.662999 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6f1448dedfa4de54eabd614cea20bd7a12db30bb599b3142b4f3befaa84637c" Mar 09 19:26:05 crc kubenswrapper[4750]: I0309 19:26:05.662464 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551406-wxvp8" Mar 09 19:26:06 crc kubenswrapper[4750]: I0309 19:26:06.146683 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551400-2jbpn"] Mar 09 19:26:06 crc kubenswrapper[4750]: I0309 19:26:06.156732 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551400-2jbpn"] Mar 09 19:26:07 crc kubenswrapper[4750]: I0309 19:26:07.390040 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78e39153-00de-4622-a05a-1b92ecac2777" path="/var/lib/kubelet/pods/78e39153-00de-4622-a05a-1b92ecac2777/volumes" Mar 09 19:26:14 crc kubenswrapper[4750]: I0309 19:26:14.373970 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:26:14 crc kubenswrapper[4750]: E0309 19:26:14.374906 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:26:21 crc kubenswrapper[4750]: I0309 19:26:21.743593 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:26:21 crc kubenswrapper[4750]: I0309 19:26:21.744623 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:26:25 crc kubenswrapper[4750]: I0309 19:26:25.374498 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:26:25 crc kubenswrapper[4750]: E0309 19:26:25.375449 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:26:36 crc kubenswrapper[4750]: I0309 19:26:36.374052 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:26:36 crc kubenswrapper[4750]: E0309 19:26:36.375505 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:26:46 crc kubenswrapper[4750]: I0309 19:26:46.934998 4750 scope.go:117] "RemoveContainer" containerID="512ea9bad820cb6cd92ae8172e62dcae900c25448f4197e1479dd6ec609e1e7d" Mar 09 19:26:51 crc kubenswrapper[4750]: I0309 19:26:51.374268 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:26:51 crc kubenswrapper[4750]: E0309 19:26:51.375320 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:26:51 crc kubenswrapper[4750]: I0309 19:26:51.743771 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:26:51 crc kubenswrapper[4750]: I0309 19:26:51.743859 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:26:51 crc kubenswrapper[4750]: I0309 19:26:51.743919 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:26:51 crc kubenswrapper[4750]: I0309 19:26:51.744741 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09feee34ec58900b16a230485b577a5b4f6613c80d8a21dc529b58c82458c699"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:26:51 crc kubenswrapper[4750]: I0309 19:26:51.744845 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://09feee34ec58900b16a230485b577a5b4f6613c80d8a21dc529b58c82458c699" gracePeriod=600 Mar 09 19:26:52 crc kubenswrapper[4750]: I0309 19:26:52.284957 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="09feee34ec58900b16a230485b577a5b4f6613c80d8a21dc529b58c82458c699" exitCode=0 Mar 09 19:26:52 crc kubenswrapper[4750]: I0309 19:26:52.285039 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"09feee34ec58900b16a230485b577a5b4f6613c80d8a21dc529b58c82458c699"} Mar 09 19:26:52 crc kubenswrapper[4750]: I0309 19:26:52.285412 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca"} Mar 09 19:26:52 crc kubenswrapper[4750]: I0309 19:26:52.285433 4750 scope.go:117] "RemoveContainer" containerID="9975b1c1f83556f2c277b325c7e1c017ec726befd861d4b2aac0e758e696960d" Mar 09 19:27:04 crc kubenswrapper[4750]: I0309 19:27:04.373438 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:27:04 crc kubenswrapper[4750]: E0309 19:27:04.374429 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:27:15 crc kubenswrapper[4750]: I0309 19:27:15.378139 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:27:15 crc kubenswrapper[4750]: E0309 19:27:15.379134 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:27:26 crc kubenswrapper[4750]: I0309 19:27:26.373484 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:27:26 crc kubenswrapper[4750]: E0309 19:27:26.374472 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:27:40 crc kubenswrapper[4750]: I0309 19:27:40.374073 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:27:40 crc kubenswrapper[4750]: E0309 19:27:40.375041 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:27:52 crc kubenswrapper[4750]: I0309 19:27:52.373581 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:27:52 crc kubenswrapper[4750]: E0309 19:27:52.374214 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.180433 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551408-w68n2"] Mar 09 19:28:00 crc kubenswrapper[4750]: E0309 19:28:00.181835 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e33970-767e-40c7-a5ba-11ad5473b535" containerName="oc" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.181858 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e33970-767e-40c7-a5ba-11ad5473b535" containerName="oc" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.182260 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8e33970-767e-40c7-a5ba-11ad5473b535" containerName="oc" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.183572 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551408-w68n2" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.187558 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.187575 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.187585 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.193622 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551408-w68n2"] Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.277762 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcsc9\" (UniqueName: \"kubernetes.io/projected/f403d5be-84de-48f4-86cf-20d75df0812e-kube-api-access-bcsc9\") pod \"auto-csr-approver-29551408-w68n2\" (UID: \"f403d5be-84de-48f4-86cf-20d75df0812e\") " pod="openshift-infra/auto-csr-approver-29551408-w68n2" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.380303 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcsc9\" (UniqueName: \"kubernetes.io/projected/f403d5be-84de-48f4-86cf-20d75df0812e-kube-api-access-bcsc9\") pod \"auto-csr-approver-29551408-w68n2\" (UID: \"f403d5be-84de-48f4-86cf-20d75df0812e\") " pod="openshift-infra/auto-csr-approver-29551408-w68n2" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.406561 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcsc9\" (UniqueName: \"kubernetes.io/projected/f403d5be-84de-48f4-86cf-20d75df0812e-kube-api-access-bcsc9\") pod \"auto-csr-approver-29551408-w68n2\" (UID: \"f403d5be-84de-48f4-86cf-20d75df0812e\") " pod="openshift-infra/auto-csr-approver-29551408-w68n2" Mar 09 19:28:00 crc kubenswrapper[4750]: I0309 19:28:00.525728 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551408-w68n2" Mar 09 19:28:01 crc kubenswrapper[4750]: I0309 19:28:01.071200 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551408-w68n2"] Mar 09 19:28:01 crc kubenswrapper[4750]: I0309 19:28:01.139777 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551408-w68n2" event={"ID":"f403d5be-84de-48f4-86cf-20d75df0812e","Type":"ContainerStarted","Data":"fbde7dea043ed8a3d6f4122126a446702ab29d611dc82dae12127dc0ac217988"} Mar 09 19:28:03 crc kubenswrapper[4750]: I0309 19:28:03.166816 4750 generic.go:334] "Generic (PLEG): container finished" podID="f403d5be-84de-48f4-86cf-20d75df0812e" containerID="d5948cb9186e92dfbb1bc934583b7f7691b869ef604f22449a9a478180f7820f" exitCode=0 Mar 09 19:28:03 crc kubenswrapper[4750]: I0309 19:28:03.166890 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551408-w68n2" event={"ID":"f403d5be-84de-48f4-86cf-20d75df0812e","Type":"ContainerDied","Data":"d5948cb9186e92dfbb1bc934583b7f7691b869ef604f22449a9a478180f7820f"} Mar 09 19:28:04 crc kubenswrapper[4750]: I0309 19:28:04.373740 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:28:04 crc kubenswrapper[4750]: E0309 19:28:04.374417 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:28:04 crc kubenswrapper[4750]: I0309 19:28:04.484847 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551408-w68n2" Mar 09 19:28:04 crc kubenswrapper[4750]: I0309 19:28:04.578358 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcsc9\" (UniqueName: \"kubernetes.io/projected/f403d5be-84de-48f4-86cf-20d75df0812e-kube-api-access-bcsc9\") pod \"f403d5be-84de-48f4-86cf-20d75df0812e\" (UID: \"f403d5be-84de-48f4-86cf-20d75df0812e\") " Mar 09 19:28:04 crc kubenswrapper[4750]: I0309 19:28:04.585877 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f403d5be-84de-48f4-86cf-20d75df0812e-kube-api-access-bcsc9" (OuterVolumeSpecName: "kube-api-access-bcsc9") pod "f403d5be-84de-48f4-86cf-20d75df0812e" (UID: "f403d5be-84de-48f4-86cf-20d75df0812e"). InnerVolumeSpecName "kube-api-access-bcsc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:28:04 crc kubenswrapper[4750]: I0309 19:28:04.680865 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcsc9\" (UniqueName: \"kubernetes.io/projected/f403d5be-84de-48f4-86cf-20d75df0812e-kube-api-access-bcsc9\") on node \"crc\" DevicePath \"\"" Mar 09 19:28:05 crc kubenswrapper[4750]: I0309 19:28:05.200324 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551408-w68n2" event={"ID":"f403d5be-84de-48f4-86cf-20d75df0812e","Type":"ContainerDied","Data":"fbde7dea043ed8a3d6f4122126a446702ab29d611dc82dae12127dc0ac217988"} Mar 09 19:28:05 crc kubenswrapper[4750]: I0309 19:28:05.200386 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbde7dea043ed8a3d6f4122126a446702ab29d611dc82dae12127dc0ac217988" Mar 09 19:28:05 crc kubenswrapper[4750]: I0309 19:28:05.200446 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551408-w68n2" Mar 09 19:28:05 crc kubenswrapper[4750]: I0309 19:28:05.578571 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551402-s4dt4"] Mar 09 19:28:05 crc kubenswrapper[4750]: I0309 19:28:05.587702 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551402-s4dt4"] Mar 09 19:28:07 crc kubenswrapper[4750]: I0309 19:28:07.388281 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9777e2-f617-4509-9521-0c73b21737af" path="/var/lib/kubelet/pods/ee9777e2-f617-4509-9521-0c73b21737af/volumes" Mar 09 19:28:19 crc kubenswrapper[4750]: I0309 19:28:19.387103 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:28:19 crc kubenswrapper[4750]: E0309 19:28:19.387950 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:28:33 crc kubenswrapper[4750]: I0309 19:28:33.373190 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:28:33 crc kubenswrapper[4750]: E0309 19:28:33.374083 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:28:45 crc kubenswrapper[4750]: I0309 19:28:45.375541 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:28:45 crc kubenswrapper[4750]: E0309 19:28:45.377051 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:28:47 crc kubenswrapper[4750]: I0309 19:28:47.094880 4750 scope.go:117] "RemoveContainer" containerID="d760fe8f5adf08f53a4dc21ee86ac830e427fc8d1f5d8e1c64d84681c2f1a893" Mar 09 19:28:57 crc kubenswrapper[4750]: I0309 19:28:57.378386 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:28:57 crc kubenswrapper[4750]: E0309 19:28:57.379152 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:29:10 crc kubenswrapper[4750]: I0309 19:29:10.373892 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:29:10 crc kubenswrapper[4750]: E0309 19:29:10.377209 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:29:21 crc kubenswrapper[4750]: I0309 19:29:21.375333 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:29:21 crc kubenswrapper[4750]: E0309 19:29:21.376138 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:29:21 crc kubenswrapper[4750]: I0309 19:29:21.743156 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:29:21 crc kubenswrapper[4750]: I0309 19:29:21.743215 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:29:32 crc kubenswrapper[4750]: I0309 19:29:32.373264 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:29:32 crc kubenswrapper[4750]: E0309 19:29:32.374040 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:29:44 crc kubenswrapper[4750]: I0309 19:29:44.372985 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:29:44 crc kubenswrapper[4750]: E0309 19:29:44.374043 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:29:51 crc kubenswrapper[4750]: I0309 19:29:51.743154 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:29:51 crc kubenswrapper[4750]: I0309 19:29:51.743857 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:29:57 crc kubenswrapper[4750]: I0309 19:29:57.374408 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:29:57 crc kubenswrapper[4750]: E0309 19:29:57.375387 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.174245 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551410-5jvnp"] Mar 09 19:30:00 crc kubenswrapper[4750]: E0309 19:30:00.175365 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f403d5be-84de-48f4-86cf-20d75df0812e" containerName="oc" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.175380 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f403d5be-84de-48f4-86cf-20d75df0812e" containerName="oc" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.175605 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f403d5be-84de-48f4-86cf-20d75df0812e" containerName="oc" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.176385 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551410-5jvnp" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.180212 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.181102 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.181497 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.204418 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhxwm\" (UniqueName: \"kubernetes.io/projected/bcf1b77f-8234-40b8-90be-b073864b5514-kube-api-access-hhxwm\") pod \"auto-csr-approver-29551410-5jvnp\" (UID: \"bcf1b77f-8234-40b8-90be-b073864b5514\") " pod="openshift-infra/auto-csr-approver-29551410-5jvnp" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.204458 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551410-5jvnp"] Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.265651 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq"] Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.267366 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.271411 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.274785 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.284341 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq"] Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.305568 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-config-volume\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.305684 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-secret-volume\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.305720 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh4dm\" (UniqueName: \"kubernetes.io/projected/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-kube-api-access-gh4dm\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.305785 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhxwm\" (UniqueName: \"kubernetes.io/projected/bcf1b77f-8234-40b8-90be-b073864b5514-kube-api-access-hhxwm\") pod \"auto-csr-approver-29551410-5jvnp\" (UID: \"bcf1b77f-8234-40b8-90be-b073864b5514\") " pod="openshift-infra/auto-csr-approver-29551410-5jvnp" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.324963 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhxwm\" (UniqueName: \"kubernetes.io/projected/bcf1b77f-8234-40b8-90be-b073864b5514-kube-api-access-hhxwm\") pod \"auto-csr-approver-29551410-5jvnp\" (UID: \"bcf1b77f-8234-40b8-90be-b073864b5514\") " pod="openshift-infra/auto-csr-approver-29551410-5jvnp" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.407465 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-config-volume\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.407841 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-secret-volume\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.408327 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-config-volume\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.408443 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh4dm\" (UniqueName: \"kubernetes.io/projected/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-kube-api-access-gh4dm\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.412507 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-secret-volume\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.440249 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh4dm\" (UniqueName: \"kubernetes.io/projected/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-kube-api-access-gh4dm\") pod \"collect-profiles-29551410-s8rcq\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.499216 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551410-5jvnp" Mar 09 19:30:00 crc kubenswrapper[4750]: I0309 19:30:00.584623 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:01 crc kubenswrapper[4750]: I0309 19:30:01.000432 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551410-5jvnp"] Mar 09 19:30:01 crc kubenswrapper[4750]: W0309 19:30:01.140852 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b98e5a7_bfd5_40e7_b35a_4b6e2649bcf4.slice/crio-f6ccf0524e93d7d9b267e57b9699b7e60b59bc1ffb28c54e8a88ecf9f4bd73bd WatchSource:0}: Error finding container f6ccf0524e93d7d9b267e57b9699b7e60b59bc1ffb28c54e8a88ecf9f4bd73bd: Status 404 returned error can't find the container with id f6ccf0524e93d7d9b267e57b9699b7e60b59bc1ffb28c54e8a88ecf9f4bd73bd Mar 09 19:30:01 crc kubenswrapper[4750]: I0309 19:30:01.151329 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq"] Mar 09 19:30:01 crc kubenswrapper[4750]: I0309 19:30:01.547516 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" event={"ID":"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4","Type":"ContainerStarted","Data":"7fd159ad10b8a8cba3952ad573df40441b97914c0b7368c37d12150131783ec6"} Mar 09 19:30:01 crc kubenswrapper[4750]: I0309 19:30:01.547780 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" event={"ID":"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4","Type":"ContainerStarted","Data":"f6ccf0524e93d7d9b267e57b9699b7e60b59bc1ffb28c54e8a88ecf9f4bd73bd"} Mar 09 19:30:01 crc kubenswrapper[4750]: I0309 19:30:01.548921 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551410-5jvnp" event={"ID":"bcf1b77f-8234-40b8-90be-b073864b5514","Type":"ContainerStarted","Data":"9e8703605b5421aa3d8c756149afe4d27ad480a58a63d1200b8af6d8736bd5d4"} Mar 09 19:30:01 crc kubenswrapper[4750]: I0309 19:30:01.577698 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" podStartSLOduration=1.577674235 podStartE2EDuration="1.577674235s" podCreationTimestamp="2026-03-09 19:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 19:30:01.568616549 +0000 UTC m=+3882.911088957" watchObservedRunningTime="2026-03-09 19:30:01.577674235 +0000 UTC m=+3882.920146643" Mar 09 19:30:02 crc kubenswrapper[4750]: I0309 19:30:02.584782 4750 generic.go:334] "Generic (PLEG): container finished" podID="2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" containerID="7fd159ad10b8a8cba3952ad573df40441b97914c0b7368c37d12150131783ec6" exitCode=0 Mar 09 19:30:02 crc kubenswrapper[4750]: I0309 19:30:02.585116 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" event={"ID":"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4","Type":"ContainerDied","Data":"7fd159ad10b8a8cba3952ad573df40441b97914c0b7368c37d12150131783ec6"} Mar 09 19:30:03 crc kubenswrapper[4750]: I0309 19:30:03.599786 4750 generic.go:334] "Generic (PLEG): container finished" podID="bcf1b77f-8234-40b8-90be-b073864b5514" containerID="40d46426df3ceea5a8bd1803aad6998d54174a2a8f861dd3f6539be030b2443d" exitCode=0 Mar 09 19:30:03 crc kubenswrapper[4750]: I0309 19:30:03.599864 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551410-5jvnp" event={"ID":"bcf1b77f-8234-40b8-90be-b073864b5514","Type":"ContainerDied","Data":"40d46426df3ceea5a8bd1803aad6998d54174a2a8f861dd3f6539be030b2443d"} Mar 09 19:30:03 crc kubenswrapper[4750]: I0309 19:30:03.970914 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.022519 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh4dm\" (UniqueName: \"kubernetes.io/projected/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-kube-api-access-gh4dm\") pod \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.022686 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-secret-volume\") pod \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.022739 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-config-volume\") pod \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\" (UID: \"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4\") " Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.023965 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-config-volume" (OuterVolumeSpecName: "config-volume") pod "2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" (UID: "2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.029409 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-kube-api-access-gh4dm" (OuterVolumeSpecName: "kube-api-access-gh4dm") pod "2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" (UID: "2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4"). InnerVolumeSpecName "kube-api-access-gh4dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.029800 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" (UID: "2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.125162 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh4dm\" (UniqueName: \"kubernetes.io/projected/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-kube-api-access-gh4dm\") on node \"crc\" DevicePath \"\"" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.125192 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.125202 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.620166 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" event={"ID":"2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4","Type":"ContainerDied","Data":"f6ccf0524e93d7d9b267e57b9699b7e60b59bc1ffb28c54e8a88ecf9f4bd73bd"} Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.620524 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6ccf0524e93d7d9b267e57b9699b7e60b59bc1ffb28c54e8a88ecf9f4bd73bd" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.620209 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq" Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.650429 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6"] Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.659400 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551365-xklq6"] Mar 09 19:30:04 crc kubenswrapper[4750]: I0309 19:30:04.991076 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551410-5jvnp" Mar 09 19:30:05 crc kubenswrapper[4750]: I0309 19:30:05.043859 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhxwm\" (UniqueName: \"kubernetes.io/projected/bcf1b77f-8234-40b8-90be-b073864b5514-kube-api-access-hhxwm\") pod \"bcf1b77f-8234-40b8-90be-b073864b5514\" (UID: \"bcf1b77f-8234-40b8-90be-b073864b5514\") " Mar 09 19:30:05 crc kubenswrapper[4750]: I0309 19:30:05.052956 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcf1b77f-8234-40b8-90be-b073864b5514-kube-api-access-hhxwm" (OuterVolumeSpecName: "kube-api-access-hhxwm") pod "bcf1b77f-8234-40b8-90be-b073864b5514" (UID: "bcf1b77f-8234-40b8-90be-b073864b5514"). InnerVolumeSpecName "kube-api-access-hhxwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:30:05 crc kubenswrapper[4750]: I0309 19:30:05.145919 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhxwm\" (UniqueName: \"kubernetes.io/projected/bcf1b77f-8234-40b8-90be-b073864b5514-kube-api-access-hhxwm\") on node \"crc\" DevicePath \"\"" Mar 09 19:30:05 crc kubenswrapper[4750]: I0309 19:30:05.389349 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53eb215e-071b-49e8-b1c7-a97471da3fa2" path="/var/lib/kubelet/pods/53eb215e-071b-49e8-b1c7-a97471da3fa2/volumes" Mar 09 19:30:05 crc kubenswrapper[4750]: I0309 19:30:05.633068 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551410-5jvnp" event={"ID":"bcf1b77f-8234-40b8-90be-b073864b5514","Type":"ContainerDied","Data":"9e8703605b5421aa3d8c756149afe4d27ad480a58a63d1200b8af6d8736bd5d4"} Mar 09 19:30:05 crc kubenswrapper[4750]: I0309 19:30:05.633117 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e8703605b5421aa3d8c756149afe4d27ad480a58a63d1200b8af6d8736bd5d4" Mar 09 19:30:05 crc kubenswrapper[4750]: I0309 19:30:05.633160 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551410-5jvnp" Mar 09 19:30:06 crc kubenswrapper[4750]: I0309 19:30:06.047386 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551404-7pwhk"] Mar 09 19:30:06 crc kubenswrapper[4750]: I0309 19:30:06.055558 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551404-7pwhk"] Mar 09 19:30:07 crc kubenswrapper[4750]: I0309 19:30:07.384386 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1523b29c-ce7d-45fc-88fd-4192c9dc2e28" path="/var/lib/kubelet/pods/1523b29c-ce7d-45fc-88fd-4192c9dc2e28/volumes" Mar 09 19:30:11 crc kubenswrapper[4750]: I0309 19:30:11.373620 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:30:12 crc kubenswrapper[4750]: I0309 19:30:12.720484 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205"} Mar 09 19:30:14 crc kubenswrapper[4750]: I0309 19:30:14.750148 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" exitCode=1 Mar 09 19:30:14 crc kubenswrapper[4750]: I0309 19:30:14.750216 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205"} Mar 09 19:30:14 crc kubenswrapper[4750]: I0309 19:30:14.750255 4750 scope.go:117] "RemoveContainer" containerID="270de6b9dedbb2b7f0b5663bc05d864098d63fb3b52d5930307c21ad31bb1c44" Mar 09 19:30:14 crc kubenswrapper[4750]: I0309 19:30:14.751342 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:30:14 crc kubenswrapper[4750]: E0309 19:30:14.752030 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:30:15 crc kubenswrapper[4750]: I0309 19:30:15.189458 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:30:15 crc kubenswrapper[4750]: I0309 19:30:15.189521 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:30:15 crc kubenswrapper[4750]: I0309 19:30:15.768842 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:30:15 crc kubenswrapper[4750]: E0309 19:30:15.770854 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:30:17 crc kubenswrapper[4750]: I0309 19:30:17.190356 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:30:17 crc kubenswrapper[4750]: I0309 19:30:17.191180 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:30:17 crc kubenswrapper[4750]: E0309 19:30:17.191409 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:30:21 crc kubenswrapper[4750]: I0309 19:30:21.743715 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:30:21 crc kubenswrapper[4750]: I0309 19:30:21.744452 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:30:21 crc kubenswrapper[4750]: I0309 19:30:21.744529 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:30:21 crc kubenswrapper[4750]: I0309 19:30:21.745704 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:30:21 crc kubenswrapper[4750]: I0309 19:30:21.745798 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" gracePeriod=600 Mar 09 19:30:21 crc kubenswrapper[4750]: E0309 19:30:21.875439 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:30:22 crc kubenswrapper[4750]: I0309 19:30:22.953341 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" exitCode=0 Mar 09 19:30:22 crc kubenswrapper[4750]: I0309 19:30:22.953445 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca"} Mar 09 19:30:22 crc kubenswrapper[4750]: I0309 19:30:22.953490 4750 scope.go:117] "RemoveContainer" containerID="09feee34ec58900b16a230485b577a5b4f6613c80d8a21dc529b58c82458c699" Mar 09 19:30:22 crc kubenswrapper[4750]: I0309 19:30:22.954423 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:30:22 crc kubenswrapper[4750]: E0309 19:30:22.954915 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:30:29 crc kubenswrapper[4750]: I0309 19:30:29.382218 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:30:29 crc kubenswrapper[4750]: E0309 19:30:29.383040 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:30:35 crc kubenswrapper[4750]: I0309 19:30:35.375223 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:30:35 crc kubenswrapper[4750]: E0309 19:30:35.376449 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:30:41 crc kubenswrapper[4750]: I0309 19:30:41.374281 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:30:41 crc kubenswrapper[4750]: E0309 19:30:41.376026 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:30:47 crc kubenswrapper[4750]: I0309 19:30:47.211710 4750 scope.go:117] "RemoveContainer" containerID="9d6bbc169c6d6e063ea2e716563a12ac31e552479910608141017e9225fa94ab" Mar 09 19:30:47 crc kubenswrapper[4750]: I0309 19:30:47.244746 4750 scope.go:117] "RemoveContainer" containerID="699a78d92c455e1d9a61a1e084d6c2515943ed9179e6f8403fa5c782d36433db" Mar 09 19:30:50 crc kubenswrapper[4750]: I0309 19:30:50.374063 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:30:50 crc kubenswrapper[4750]: E0309 19:30:50.375378 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:30:54 crc kubenswrapper[4750]: I0309 19:30:54.374406 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:30:54 crc kubenswrapper[4750]: E0309 19:30:54.378474 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:31:04 crc kubenswrapper[4750]: I0309 19:31:04.374302 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:31:04 crc kubenswrapper[4750]: E0309 19:31:04.375062 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:31:05 crc kubenswrapper[4750]: I0309 19:31:05.374179 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:31:05 crc kubenswrapper[4750]: E0309 19:31:05.374815 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:31:15 crc kubenswrapper[4750]: I0309 19:31:15.374733 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:31:15 crc kubenswrapper[4750]: E0309 19:31:15.376221 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:31:16 crc kubenswrapper[4750]: I0309 19:31:16.373458 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:31:16 crc kubenswrapper[4750]: E0309 19:31:16.374055 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:31:30 crc kubenswrapper[4750]: I0309 19:31:30.373944 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:31:30 crc kubenswrapper[4750]: I0309 19:31:30.374759 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:31:30 crc kubenswrapper[4750]: E0309 19:31:30.375126 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:31:30 crc kubenswrapper[4750]: E0309 19:31:30.375357 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:31:45 crc kubenswrapper[4750]: I0309 19:31:45.373558 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:31:45 crc kubenswrapper[4750]: I0309 19:31:45.374151 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:31:45 crc kubenswrapper[4750]: E0309 19:31:45.374274 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:31:45 crc kubenswrapper[4750]: E0309 19:31:45.374457 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:31:56 crc kubenswrapper[4750]: I0309 19:31:56.373593 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:31:56 crc kubenswrapper[4750]: E0309 19:31:56.374722 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.442516 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l5gg2"] Mar 09 19:31:57 crc kubenswrapper[4750]: E0309 19:31:57.443139 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" containerName="collect-profiles" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.443157 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" containerName="collect-profiles" Mar 09 19:31:57 crc kubenswrapper[4750]: E0309 19:31:57.443194 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcf1b77f-8234-40b8-90be-b073864b5514" containerName="oc" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.443203 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcf1b77f-8234-40b8-90be-b073864b5514" containerName="oc" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.443474 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcf1b77f-8234-40b8-90be-b073864b5514" containerName="oc" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.443497 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" containerName="collect-profiles" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.445422 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.456371 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5gg2"] Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.579001 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk2zt\" (UniqueName: \"kubernetes.io/projected/d2405775-9665-47e8-ba53-7912203e55bc-kube-api-access-hk2zt\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.579070 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-catalog-content\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.579226 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-utilities\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.680866 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk2zt\" (UniqueName: \"kubernetes.io/projected/d2405775-9665-47e8-ba53-7912203e55bc-kube-api-access-hk2zt\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.680920 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-catalog-content\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.681009 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-utilities\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.681464 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-utilities\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.681573 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-catalog-content\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.700286 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk2zt\" (UniqueName: \"kubernetes.io/projected/d2405775-9665-47e8-ba53-7912203e55bc-kube-api-access-hk2zt\") pod \"redhat-operators-l5gg2\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:57 crc kubenswrapper[4750]: I0309 19:31:57.764879 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:31:58 crc kubenswrapper[4750]: I0309 19:31:58.268441 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5gg2"] Mar 09 19:31:59 crc kubenswrapper[4750]: I0309 19:31:59.201111 4750 generic.go:334] "Generic (PLEG): container finished" podID="d2405775-9665-47e8-ba53-7912203e55bc" containerID="b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0" exitCode=0 Mar 09 19:31:59 crc kubenswrapper[4750]: I0309 19:31:59.201239 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5gg2" event={"ID":"d2405775-9665-47e8-ba53-7912203e55bc","Type":"ContainerDied","Data":"b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0"} Mar 09 19:31:59 crc kubenswrapper[4750]: I0309 19:31:59.201489 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5gg2" event={"ID":"d2405775-9665-47e8-ba53-7912203e55bc","Type":"ContainerStarted","Data":"4a881b5658ecc09d9cbd912fcfbfa8c3a2aa9616d63a1259769740ea48c0d26c"} Mar 09 19:31:59 crc kubenswrapper[4750]: I0309 19:31:59.204464 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:31:59 crc kubenswrapper[4750]: I0309 19:31:59.387717 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:31:59 crc kubenswrapper[4750]: E0309 19:31:59.388196 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.149851 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551412-4qkjs"] Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.151939 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551412-4qkjs" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.154247 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.154758 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.154951 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.171617 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551412-4qkjs"] Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.215001 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5gg2" event={"ID":"d2405775-9665-47e8-ba53-7912203e55bc","Type":"ContainerStarted","Data":"69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb"} Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.263247 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwvkd\" (UniqueName: \"kubernetes.io/projected/c5cdd0e0-e091-4a05-bab5-872aeef5d293-kube-api-access-rwvkd\") pod \"auto-csr-approver-29551412-4qkjs\" (UID: \"c5cdd0e0-e091-4a05-bab5-872aeef5d293\") " pod="openshift-infra/auto-csr-approver-29551412-4qkjs" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.365650 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwvkd\" (UniqueName: \"kubernetes.io/projected/c5cdd0e0-e091-4a05-bab5-872aeef5d293-kube-api-access-rwvkd\") pod \"auto-csr-approver-29551412-4qkjs\" (UID: \"c5cdd0e0-e091-4a05-bab5-872aeef5d293\") " pod="openshift-infra/auto-csr-approver-29551412-4qkjs" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.390231 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwvkd\" (UniqueName: \"kubernetes.io/projected/c5cdd0e0-e091-4a05-bab5-872aeef5d293-kube-api-access-rwvkd\") pod \"auto-csr-approver-29551412-4qkjs\" (UID: \"c5cdd0e0-e091-4a05-bab5-872aeef5d293\") " pod="openshift-infra/auto-csr-approver-29551412-4qkjs" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.474462 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551412-4qkjs" Mar 09 19:32:00 crc kubenswrapper[4750]: I0309 19:32:00.963782 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551412-4qkjs"] Mar 09 19:32:00 crc kubenswrapper[4750]: W0309 19:32:00.970412 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5cdd0e0_e091_4a05_bab5_872aeef5d293.slice/crio-49fc640a896df079d0eb8b04ec67dfc761fef557cd46b35c564210322c1f2fa6 WatchSource:0}: Error finding container 49fc640a896df079d0eb8b04ec67dfc761fef557cd46b35c564210322c1f2fa6: Status 404 returned error can't find the container with id 49fc640a896df079d0eb8b04ec67dfc761fef557cd46b35c564210322c1f2fa6 Mar 09 19:32:01 crc kubenswrapper[4750]: I0309 19:32:01.227279 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551412-4qkjs" event={"ID":"c5cdd0e0-e091-4a05-bab5-872aeef5d293","Type":"ContainerStarted","Data":"49fc640a896df079d0eb8b04ec67dfc761fef557cd46b35c564210322c1f2fa6"} Mar 09 19:32:03 crc kubenswrapper[4750]: I0309 19:32:03.253601 4750 generic.go:334] "Generic (PLEG): container finished" podID="c5cdd0e0-e091-4a05-bab5-872aeef5d293" containerID="7d5be83a6dfd3a19ced9532ac56dc14e399874126bc63a039029797312850cb7" exitCode=0 Mar 09 19:32:03 crc kubenswrapper[4750]: I0309 19:32:03.253684 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551412-4qkjs" event={"ID":"c5cdd0e0-e091-4a05-bab5-872aeef5d293","Type":"ContainerDied","Data":"7d5be83a6dfd3a19ced9532ac56dc14e399874126bc63a039029797312850cb7"} Mar 09 19:32:03 crc kubenswrapper[4750]: I0309 19:32:03.258040 4750 generic.go:334] "Generic (PLEG): container finished" podID="d2405775-9665-47e8-ba53-7912203e55bc" containerID="69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb" exitCode=0 Mar 09 19:32:03 crc kubenswrapper[4750]: I0309 19:32:03.258086 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5gg2" event={"ID":"d2405775-9665-47e8-ba53-7912203e55bc","Type":"ContainerDied","Data":"69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb"} Mar 09 19:32:04 crc kubenswrapper[4750]: I0309 19:32:04.271735 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5gg2" event={"ID":"d2405775-9665-47e8-ba53-7912203e55bc","Type":"ContainerStarted","Data":"6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5"} Mar 09 19:32:04 crc kubenswrapper[4750]: I0309 19:32:04.313469 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l5gg2" podStartSLOduration=2.6514622770000003 podStartE2EDuration="7.313444291s" podCreationTimestamp="2026-03-09 19:31:57 +0000 UTC" firstStartedPulling="2026-03-09 19:31:59.204207921 +0000 UTC m=+4000.546680319" lastFinishedPulling="2026-03-09 19:32:03.866189925 +0000 UTC m=+4005.208662333" observedRunningTime="2026-03-09 19:32:04.295458325 +0000 UTC m=+4005.637930753" watchObservedRunningTime="2026-03-09 19:32:04.313444291 +0000 UTC m=+4005.655916709" Mar 09 19:32:05 crc kubenswrapper[4750]: I0309 19:32:05.280936 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551412-4qkjs" event={"ID":"c5cdd0e0-e091-4a05-bab5-872aeef5d293","Type":"ContainerDied","Data":"49fc640a896df079d0eb8b04ec67dfc761fef557cd46b35c564210322c1f2fa6"} Mar 09 19:32:05 crc kubenswrapper[4750]: I0309 19:32:05.281190 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49fc640a896df079d0eb8b04ec67dfc761fef557cd46b35c564210322c1f2fa6" Mar 09 19:32:05 crc kubenswrapper[4750]: I0309 19:32:05.286476 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551412-4qkjs" Mar 09 19:32:05 crc kubenswrapper[4750]: I0309 19:32:05.488622 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwvkd\" (UniqueName: \"kubernetes.io/projected/c5cdd0e0-e091-4a05-bab5-872aeef5d293-kube-api-access-rwvkd\") pod \"c5cdd0e0-e091-4a05-bab5-872aeef5d293\" (UID: \"c5cdd0e0-e091-4a05-bab5-872aeef5d293\") " Mar 09 19:32:05 crc kubenswrapper[4750]: I0309 19:32:05.499984 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cdd0e0-e091-4a05-bab5-872aeef5d293-kube-api-access-rwvkd" (OuterVolumeSpecName: "kube-api-access-rwvkd") pod "c5cdd0e0-e091-4a05-bab5-872aeef5d293" (UID: "c5cdd0e0-e091-4a05-bab5-872aeef5d293"). InnerVolumeSpecName "kube-api-access-rwvkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:32:05 crc kubenswrapper[4750]: I0309 19:32:05.591757 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwvkd\" (UniqueName: \"kubernetes.io/projected/c5cdd0e0-e091-4a05-bab5-872aeef5d293-kube-api-access-rwvkd\") on node \"crc\" DevicePath \"\"" Mar 09 19:32:06 crc kubenswrapper[4750]: I0309 19:32:06.305250 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551412-4qkjs" Mar 09 19:32:06 crc kubenswrapper[4750]: I0309 19:32:06.392108 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551406-wxvp8"] Mar 09 19:32:06 crc kubenswrapper[4750]: I0309 19:32:06.409470 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551406-wxvp8"] Mar 09 19:32:07 crc kubenswrapper[4750]: I0309 19:32:07.386922 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8e33970-767e-40c7-a5ba-11ad5473b535" path="/var/lib/kubelet/pods/e8e33970-767e-40c7-a5ba-11ad5473b535/volumes" Mar 09 19:32:07 crc kubenswrapper[4750]: I0309 19:32:07.765145 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:32:07 crc kubenswrapper[4750]: I0309 19:32:07.765255 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:32:08 crc kubenswrapper[4750]: I0309 19:32:08.839111 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l5gg2" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="registry-server" probeResult="failure" output=< Mar 09 19:32:08 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:32:08 crc kubenswrapper[4750]: > Mar 09 19:32:10 crc kubenswrapper[4750]: I0309 19:32:10.374270 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:32:10 crc kubenswrapper[4750]: E0309 19:32:10.376230 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:32:11 crc kubenswrapper[4750]: I0309 19:32:11.374359 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:32:11 crc kubenswrapper[4750]: E0309 19:32:11.375331 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:32:17 crc kubenswrapper[4750]: I0309 19:32:17.885813 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:32:17 crc kubenswrapper[4750]: I0309 19:32:17.953043 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:32:18 crc kubenswrapper[4750]: I0309 19:32:18.157304 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5gg2"] Mar 09 19:32:19 crc kubenswrapper[4750]: I0309 19:32:19.465574 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l5gg2" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="registry-server" containerID="cri-o://6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5" gracePeriod=2 Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.019746 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.119577 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-utilities\") pod \"d2405775-9665-47e8-ba53-7912203e55bc\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.120095 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk2zt\" (UniqueName: \"kubernetes.io/projected/d2405775-9665-47e8-ba53-7912203e55bc-kube-api-access-hk2zt\") pod \"d2405775-9665-47e8-ba53-7912203e55bc\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.120146 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-catalog-content\") pod \"d2405775-9665-47e8-ba53-7912203e55bc\" (UID: \"d2405775-9665-47e8-ba53-7912203e55bc\") " Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.120407 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-utilities" (OuterVolumeSpecName: "utilities") pod "d2405775-9665-47e8-ba53-7912203e55bc" (UID: "d2405775-9665-47e8-ba53-7912203e55bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.120969 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.127165 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2405775-9665-47e8-ba53-7912203e55bc-kube-api-access-hk2zt" (OuterVolumeSpecName: "kube-api-access-hk2zt") pod "d2405775-9665-47e8-ba53-7912203e55bc" (UID: "d2405775-9665-47e8-ba53-7912203e55bc"). InnerVolumeSpecName "kube-api-access-hk2zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.225149 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk2zt\" (UniqueName: \"kubernetes.io/projected/d2405775-9665-47e8-ba53-7912203e55bc-kube-api-access-hk2zt\") on node \"crc\" DevicePath \"\"" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.247359 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2405775-9665-47e8-ba53-7912203e55bc" (UID: "d2405775-9665-47e8-ba53-7912203e55bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.327526 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2405775-9665-47e8-ba53-7912203e55bc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.493248 4750 generic.go:334] "Generic (PLEG): container finished" podID="d2405775-9665-47e8-ba53-7912203e55bc" containerID="6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5" exitCode=0 Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.493333 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5gg2" event={"ID":"d2405775-9665-47e8-ba53-7912203e55bc","Type":"ContainerDied","Data":"6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5"} Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.493378 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5gg2" event={"ID":"d2405775-9665-47e8-ba53-7912203e55bc","Type":"ContainerDied","Data":"4a881b5658ecc09d9cbd912fcfbfa8c3a2aa9616d63a1259769740ea48c0d26c"} Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.493429 4750 scope.go:117] "RemoveContainer" containerID="6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.493793 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5gg2" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.563223 4750 scope.go:117] "RemoveContainer" containerID="69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.573190 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5gg2"] Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.585372 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l5gg2"] Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.606464 4750 scope.go:117] "RemoveContainer" containerID="b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.646917 4750 scope.go:117] "RemoveContainer" containerID="6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5" Mar 09 19:32:20 crc kubenswrapper[4750]: E0309 19:32:20.647377 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5\": container with ID starting with 6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5 not found: ID does not exist" containerID="6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.647426 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5"} err="failed to get container status \"6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5\": rpc error: code = NotFound desc = could not find container \"6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5\": container with ID starting with 6ce972dbc584db4d09af789208a1070e34a5e7e32950f5f4446e6ea6d76b91d5 not found: ID does not exist" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.647459 4750 scope.go:117] "RemoveContainer" containerID="69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb" Mar 09 19:32:20 crc kubenswrapper[4750]: E0309 19:32:20.647864 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb\": container with ID starting with 69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb not found: ID does not exist" containerID="69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.647921 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb"} err="failed to get container status \"69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb\": rpc error: code = NotFound desc = could not find container \"69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb\": container with ID starting with 69c11546a9a34a79359546fb66d48d8fe96187ac2e829e54359b2d929057f8cb not found: ID does not exist" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.647954 4750 scope.go:117] "RemoveContainer" containerID="b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0" Mar 09 19:32:20 crc kubenswrapper[4750]: E0309 19:32:20.648296 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0\": container with ID starting with b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0 not found: ID does not exist" containerID="b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0" Mar 09 19:32:20 crc kubenswrapper[4750]: I0309 19:32:20.648330 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0"} err="failed to get container status \"b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0\": rpc error: code = NotFound desc = could not find container \"b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0\": container with ID starting with b434498e1e469db8e31c5ff9f1c8b0c546e887bea39723ab9cb5130018abc7f0 not found: ID does not exist" Mar 09 19:32:21 crc kubenswrapper[4750]: I0309 19:32:21.388471 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2405775-9665-47e8-ba53-7912203e55bc" path="/var/lib/kubelet/pods/d2405775-9665-47e8-ba53-7912203e55bc/volumes" Mar 09 19:32:25 crc kubenswrapper[4750]: I0309 19:32:25.378278 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:32:25 crc kubenswrapper[4750]: E0309 19:32:25.379457 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:32:26 crc kubenswrapper[4750]: I0309 19:32:26.373838 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:32:26 crc kubenswrapper[4750]: E0309 19:32:26.374585 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:32:36 crc kubenswrapper[4750]: I0309 19:32:36.373770 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:32:36 crc kubenswrapper[4750]: E0309 19:32:36.374730 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:32:38 crc kubenswrapper[4750]: I0309 19:32:38.374376 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:32:38 crc kubenswrapper[4750]: E0309 19:32:38.374972 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:32:47 crc kubenswrapper[4750]: I0309 19:32:47.462773 4750 scope.go:117] "RemoveContainer" containerID="af19d8bc0656e16d31076a3067cb291b01619330b0337ea9656f1b94fa51124e" Mar 09 19:32:51 crc kubenswrapper[4750]: I0309 19:32:51.374609 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:32:51 crc kubenswrapper[4750]: E0309 19:32:51.377005 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:32:53 crc kubenswrapper[4750]: I0309 19:32:53.375008 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:32:53 crc kubenswrapper[4750]: E0309 19:32:53.375461 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:33:06 crc kubenswrapper[4750]: I0309 19:33:06.376125 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:33:06 crc kubenswrapper[4750]: E0309 19:33:06.379448 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:33:08 crc kubenswrapper[4750]: I0309 19:33:08.374361 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:33:08 crc kubenswrapper[4750]: E0309 19:33:08.374952 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:33:18 crc kubenswrapper[4750]: I0309 19:33:18.375005 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:33:18 crc kubenswrapper[4750]: E0309 19:33:18.376085 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:33:19 crc kubenswrapper[4750]: I0309 19:33:19.388295 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:33:19 crc kubenswrapper[4750]: E0309 19:33:19.389286 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:33:31 crc kubenswrapper[4750]: I0309 19:33:31.374789 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:33:31 crc kubenswrapper[4750]: E0309 19:33:31.376027 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:33:33 crc kubenswrapper[4750]: I0309 19:33:33.373687 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:33:33 crc kubenswrapper[4750]: E0309 19:33:33.374520 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.340438 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sxg74"] Mar 09 19:33:40 crc kubenswrapper[4750]: E0309 19:33:40.341458 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="extract-utilities" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.341476 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="extract-utilities" Mar 09 19:33:40 crc kubenswrapper[4750]: E0309 19:33:40.341493 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cdd0e0-e091-4a05-bab5-872aeef5d293" containerName="oc" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.341501 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cdd0e0-e091-4a05-bab5-872aeef5d293" containerName="oc" Mar 09 19:33:40 crc kubenswrapper[4750]: E0309 19:33:40.341521 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="registry-server" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.341529 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="registry-server" Mar 09 19:33:40 crc kubenswrapper[4750]: E0309 19:33:40.341552 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="extract-content" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.341559 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="extract-content" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.341836 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2405775-9665-47e8-ba53-7912203e55bc" containerName="registry-server" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.341852 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cdd0e0-e091-4a05-bab5-872aeef5d293" containerName="oc" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.346341 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.370604 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sxg74"] Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.428257 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-catalog-content\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.428538 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-utilities\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.428755 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b24sb\" (UniqueName: \"kubernetes.io/projected/cdb69170-52a0-49f7-b25f-a5cae31003a9-kube-api-access-b24sb\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.530920 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b24sb\" (UniqueName: \"kubernetes.io/projected/cdb69170-52a0-49f7-b25f-a5cae31003a9-kube-api-access-b24sb\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.531034 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-catalog-content\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.531058 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-utilities\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.531522 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-utilities\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.531697 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-catalog-content\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.564139 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b24sb\" (UniqueName: \"kubernetes.io/projected/cdb69170-52a0-49f7-b25f-a5cae31003a9-kube-api-access-b24sb\") pod \"community-operators-sxg74\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:40 crc kubenswrapper[4750]: I0309 19:33:40.673354 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:41 crc kubenswrapper[4750]: I0309 19:33:41.218807 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sxg74"] Mar 09 19:33:41 crc kubenswrapper[4750]: I0309 19:33:41.463919 4750 generic.go:334] "Generic (PLEG): container finished" podID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerID="eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d" exitCode=0 Mar 09 19:33:41 crc kubenswrapper[4750]: I0309 19:33:41.463980 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxg74" event={"ID":"cdb69170-52a0-49f7-b25f-a5cae31003a9","Type":"ContainerDied","Data":"eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d"} Mar 09 19:33:41 crc kubenswrapper[4750]: I0309 19:33:41.464200 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxg74" event={"ID":"cdb69170-52a0-49f7-b25f-a5cae31003a9","Type":"ContainerStarted","Data":"fe2f99b00dded3040f22a9cca3128135b6c223824ce36833ac80a7498ebfe543"} Mar 09 19:33:43 crc kubenswrapper[4750]: I0309 19:33:43.488411 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxg74" event={"ID":"cdb69170-52a0-49f7-b25f-a5cae31003a9","Type":"ContainerStarted","Data":"c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540"} Mar 09 19:33:44 crc kubenswrapper[4750]: I0309 19:33:44.501722 4750 generic.go:334] "Generic (PLEG): container finished" podID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerID="c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540" exitCode=0 Mar 09 19:33:44 crc kubenswrapper[4750]: I0309 19:33:44.501809 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxg74" event={"ID":"cdb69170-52a0-49f7-b25f-a5cae31003a9","Type":"ContainerDied","Data":"c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540"} Mar 09 19:33:45 crc kubenswrapper[4750]: I0309 19:33:45.377034 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:33:45 crc kubenswrapper[4750]: E0309 19:33:45.378144 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:33:45 crc kubenswrapper[4750]: I0309 19:33:45.516190 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxg74" event={"ID":"cdb69170-52a0-49f7-b25f-a5cae31003a9","Type":"ContainerStarted","Data":"bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d"} Mar 09 19:33:45 crc kubenswrapper[4750]: I0309 19:33:45.546061 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sxg74" podStartSLOduration=2.101050755 podStartE2EDuration="5.546033046s" podCreationTimestamp="2026-03-09 19:33:40 +0000 UTC" firstStartedPulling="2026-03-09 19:33:41.465393134 +0000 UTC m=+4102.807865532" lastFinishedPulling="2026-03-09 19:33:44.910375375 +0000 UTC m=+4106.252847823" observedRunningTime="2026-03-09 19:33:45.534035995 +0000 UTC m=+4106.876508433" watchObservedRunningTime="2026-03-09 19:33:45.546033046 +0000 UTC m=+4106.888505484" Mar 09 19:33:48 crc kubenswrapper[4750]: I0309 19:33:48.373923 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:33:48 crc kubenswrapper[4750]: E0309 19:33:48.376082 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:33:50 crc kubenswrapper[4750]: I0309 19:33:50.674732 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:50 crc kubenswrapper[4750]: I0309 19:33:50.675141 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:50 crc kubenswrapper[4750]: I0309 19:33:50.748560 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:51 crc kubenswrapper[4750]: I0309 19:33:51.673663 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:51 crc kubenswrapper[4750]: I0309 19:33:51.742803 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sxg74"] Mar 09 19:33:53 crc kubenswrapper[4750]: I0309 19:33:53.600285 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sxg74" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="registry-server" containerID="cri-o://bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d" gracePeriod=2 Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.223652 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.252953 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b24sb\" (UniqueName: \"kubernetes.io/projected/cdb69170-52a0-49f7-b25f-a5cae31003a9-kube-api-access-b24sb\") pod \"cdb69170-52a0-49f7-b25f-a5cae31003a9\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.253055 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-catalog-content\") pod \"cdb69170-52a0-49f7-b25f-a5cae31003a9\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.253094 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-utilities\") pod \"cdb69170-52a0-49f7-b25f-a5cae31003a9\" (UID: \"cdb69170-52a0-49f7-b25f-a5cae31003a9\") " Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.254810 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-utilities" (OuterVolumeSpecName: "utilities") pod "cdb69170-52a0-49f7-b25f-a5cae31003a9" (UID: "cdb69170-52a0-49f7-b25f-a5cae31003a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.265246 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdb69170-52a0-49f7-b25f-a5cae31003a9-kube-api-access-b24sb" (OuterVolumeSpecName: "kube-api-access-b24sb") pod "cdb69170-52a0-49f7-b25f-a5cae31003a9" (UID: "cdb69170-52a0-49f7-b25f-a5cae31003a9"). InnerVolumeSpecName "kube-api-access-b24sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.353532 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdb69170-52a0-49f7-b25f-a5cae31003a9" (UID: "cdb69170-52a0-49f7-b25f-a5cae31003a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.355084 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b24sb\" (UniqueName: \"kubernetes.io/projected/cdb69170-52a0-49f7-b25f-a5cae31003a9-kube-api-access-b24sb\") on node \"crc\" DevicePath \"\"" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.355117 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.355127 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb69170-52a0-49f7-b25f-a5cae31003a9-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.615379 4750 generic.go:334] "Generic (PLEG): container finished" podID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerID="bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d" exitCode=0 Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.615439 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxg74" event={"ID":"cdb69170-52a0-49f7-b25f-a5cae31003a9","Type":"ContainerDied","Data":"bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d"} Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.615466 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxg74" event={"ID":"cdb69170-52a0-49f7-b25f-a5cae31003a9","Type":"ContainerDied","Data":"fe2f99b00dded3040f22a9cca3128135b6c223824ce36833ac80a7498ebfe543"} Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.615484 4750 scope.go:117] "RemoveContainer" containerID="bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.615679 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxg74" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.667800 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sxg74"] Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.675930 4750 scope.go:117] "RemoveContainer" containerID="c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.681699 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sxg74"] Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.711999 4750 scope.go:117] "RemoveContainer" containerID="eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.779562 4750 scope.go:117] "RemoveContainer" containerID="bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d" Mar 09 19:33:54 crc kubenswrapper[4750]: E0309 19:33:54.780379 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d\": container with ID starting with bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d not found: ID does not exist" containerID="bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.780426 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d"} err="failed to get container status \"bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d\": rpc error: code = NotFound desc = could not find container \"bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d\": container with ID starting with bc02a92fc3d8203ec399a63cba21015b8dda543f51e9b23579d6343500913e1d not found: ID does not exist" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.780453 4750 scope.go:117] "RemoveContainer" containerID="c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540" Mar 09 19:33:54 crc kubenswrapper[4750]: E0309 19:33:54.781035 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540\": container with ID starting with c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540 not found: ID does not exist" containerID="c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.781137 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540"} err="failed to get container status \"c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540\": rpc error: code = NotFound desc = could not find container \"c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540\": container with ID starting with c2646825a5069639e1a8bc9114ab1a0eea78e0b6f406d1ba356f6bd1dbdd5540 not found: ID does not exist" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.781210 4750 scope.go:117] "RemoveContainer" containerID="eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d" Mar 09 19:33:54 crc kubenswrapper[4750]: E0309 19:33:54.781614 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d\": container with ID starting with eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d not found: ID does not exist" containerID="eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d" Mar 09 19:33:54 crc kubenswrapper[4750]: I0309 19:33:54.781661 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d"} err="failed to get container status \"eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d\": rpc error: code = NotFound desc = could not find container \"eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d\": container with ID starting with eb4e5a1c7f26dbfd6a6c02ca40f06b7ae749733f204c61403b755c4852fd2e0d not found: ID does not exist" Mar 09 19:33:55 crc kubenswrapper[4750]: I0309 19:33:55.395118 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" path="/var/lib/kubelet/pods/cdb69170-52a0-49f7-b25f-a5cae31003a9/volumes" Mar 09 19:33:56 crc kubenswrapper[4750]: I0309 19:33:56.373746 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:33:56 crc kubenswrapper[4750]: E0309 19:33:56.374080 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.176596 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551414-qv9bj"] Mar 09 19:34:00 crc kubenswrapper[4750]: E0309 19:34:00.179542 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="extract-utilities" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.179763 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="extract-utilities" Mar 09 19:34:00 crc kubenswrapper[4750]: E0309 19:34:00.179907 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="registry-server" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.180023 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="registry-server" Mar 09 19:34:00 crc kubenswrapper[4750]: E0309 19:34:00.180173 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="extract-content" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.180307 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="extract-content" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.181026 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdb69170-52a0-49f7-b25f-a5cae31003a9" containerName="registry-server" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.182454 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.185997 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.187324 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.187366 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.200507 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551414-qv9bj"] Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.304498 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-992qj\" (UniqueName: \"kubernetes.io/projected/8c511856-8984-4c7c-8af1-3c093025dce8-kube-api-access-992qj\") pod \"auto-csr-approver-29551414-qv9bj\" (UID: \"8c511856-8984-4c7c-8af1-3c093025dce8\") " pod="openshift-infra/auto-csr-approver-29551414-qv9bj" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.374402 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:34:00 crc kubenswrapper[4750]: E0309 19:34:00.375023 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.407352 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-992qj\" (UniqueName: \"kubernetes.io/projected/8c511856-8984-4c7c-8af1-3c093025dce8-kube-api-access-992qj\") pod \"auto-csr-approver-29551414-qv9bj\" (UID: \"8c511856-8984-4c7c-8af1-3c093025dce8\") " pod="openshift-infra/auto-csr-approver-29551414-qv9bj" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.436933 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-992qj\" (UniqueName: \"kubernetes.io/projected/8c511856-8984-4c7c-8af1-3c093025dce8-kube-api-access-992qj\") pod \"auto-csr-approver-29551414-qv9bj\" (UID: \"8c511856-8984-4c7c-8af1-3c093025dce8\") " pod="openshift-infra/auto-csr-approver-29551414-qv9bj" Mar 09 19:34:00 crc kubenswrapper[4750]: I0309 19:34:00.519269 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" Mar 09 19:34:01 crc kubenswrapper[4750]: I0309 19:34:01.038066 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551414-qv9bj"] Mar 09 19:34:01 crc kubenswrapper[4750]: W0309 19:34:01.039874 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c511856_8984_4c7c_8af1_3c093025dce8.slice/crio-0acc33d99105cc08a2678ba643d86dbc6ee63e21993e0c59c4bd29edad05ba63 WatchSource:0}: Error finding container 0acc33d99105cc08a2678ba643d86dbc6ee63e21993e0c59c4bd29edad05ba63: Status 404 returned error can't find the container with id 0acc33d99105cc08a2678ba643d86dbc6ee63e21993e0c59c4bd29edad05ba63 Mar 09 19:34:01 crc kubenswrapper[4750]: I0309 19:34:01.716111 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" event={"ID":"8c511856-8984-4c7c-8af1-3c093025dce8","Type":"ContainerStarted","Data":"0acc33d99105cc08a2678ba643d86dbc6ee63e21993e0c59c4bd29edad05ba63"} Mar 09 19:34:02 crc kubenswrapper[4750]: I0309 19:34:02.726141 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" event={"ID":"8c511856-8984-4c7c-8af1-3c093025dce8","Type":"ContainerStarted","Data":"e3a8c85deb2996c38883afeddb20c941be784426a909a0d83385f4f8c00e1003"} Mar 09 19:34:02 crc kubenswrapper[4750]: I0309 19:34:02.742492 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" podStartSLOduration=1.5281949639999999 podStartE2EDuration="2.742467557s" podCreationTimestamp="2026-03-09 19:34:00 +0000 UTC" firstStartedPulling="2026-03-09 19:34:01.044924421 +0000 UTC m=+4122.387396829" lastFinishedPulling="2026-03-09 19:34:02.259197014 +0000 UTC m=+4123.601669422" observedRunningTime="2026-03-09 19:34:02.740682579 +0000 UTC m=+4124.083154997" watchObservedRunningTime="2026-03-09 19:34:02.742467557 +0000 UTC m=+4124.084939995" Mar 09 19:34:03 crc kubenswrapper[4750]: I0309 19:34:03.736528 4750 generic.go:334] "Generic (PLEG): container finished" podID="8c511856-8984-4c7c-8af1-3c093025dce8" containerID="e3a8c85deb2996c38883afeddb20c941be784426a909a0d83385f4f8c00e1003" exitCode=0 Mar 09 19:34:03 crc kubenswrapper[4750]: I0309 19:34:03.736611 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" event={"ID":"8c511856-8984-4c7c-8af1-3c093025dce8","Type":"ContainerDied","Data":"e3a8c85deb2996c38883afeddb20c941be784426a909a0d83385f4f8c00e1003"} Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.170425 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.215386 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-992qj\" (UniqueName: \"kubernetes.io/projected/8c511856-8984-4c7c-8af1-3c093025dce8-kube-api-access-992qj\") pod \"8c511856-8984-4c7c-8af1-3c093025dce8\" (UID: \"8c511856-8984-4c7c-8af1-3c093025dce8\") " Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.222530 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c511856-8984-4c7c-8af1-3c093025dce8-kube-api-access-992qj" (OuterVolumeSpecName: "kube-api-access-992qj") pod "8c511856-8984-4c7c-8af1-3c093025dce8" (UID: "8c511856-8984-4c7c-8af1-3c093025dce8"). InnerVolumeSpecName "kube-api-access-992qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.318121 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-992qj\" (UniqueName: \"kubernetes.io/projected/8c511856-8984-4c7c-8af1-3c093025dce8-kube-api-access-992qj\") on node \"crc\" DevicePath \"\"" Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.760837 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" event={"ID":"8c511856-8984-4c7c-8af1-3c093025dce8","Type":"ContainerDied","Data":"0acc33d99105cc08a2678ba643d86dbc6ee63e21993e0c59c4bd29edad05ba63"} Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.760888 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0acc33d99105cc08a2678ba643d86dbc6ee63e21993e0c59c4bd29edad05ba63" Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.760968 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551414-qv9bj" Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.821517 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551408-w68n2"] Mar 09 19:34:05 crc kubenswrapper[4750]: I0309 19:34:05.841624 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551408-w68n2"] Mar 09 19:34:07 crc kubenswrapper[4750]: I0309 19:34:07.374076 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:34:07 crc kubenswrapper[4750]: E0309 19:34:07.374799 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:34:07 crc kubenswrapper[4750]: I0309 19:34:07.387717 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f403d5be-84de-48f4-86cf-20d75df0812e" path="/var/lib/kubelet/pods/f403d5be-84de-48f4-86cf-20d75df0812e/volumes" Mar 09 19:34:13 crc kubenswrapper[4750]: I0309 19:34:13.374626 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:34:13 crc kubenswrapper[4750]: E0309 19:34:13.376009 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:34:20 crc kubenswrapper[4750]: I0309 19:34:20.374400 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:34:20 crc kubenswrapper[4750]: E0309 19:34:20.375511 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:34:28 crc kubenswrapper[4750]: I0309 19:34:28.374504 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:34:28 crc kubenswrapper[4750]: E0309 19:34:28.375803 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:34:35 crc kubenswrapper[4750]: I0309 19:34:35.374189 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:34:35 crc kubenswrapper[4750]: E0309 19:34:35.376578 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:34:39 crc kubenswrapper[4750]: I0309 19:34:39.388037 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:34:39 crc kubenswrapper[4750]: E0309 19:34:39.388919 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:34:47 crc kubenswrapper[4750]: I0309 19:34:47.373764 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:34:47 crc kubenswrapper[4750]: E0309 19:34:47.375129 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:34:47 crc kubenswrapper[4750]: I0309 19:34:47.611024 4750 scope.go:117] "RemoveContainer" containerID="d5948cb9186e92dfbb1bc934583b7f7691b869ef604f22449a9a478180f7820f" Mar 09 19:34:51 crc kubenswrapper[4750]: I0309 19:34:51.375972 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:34:51 crc kubenswrapper[4750]: E0309 19:34:51.377091 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:35:00 crc kubenswrapper[4750]: I0309 19:35:00.375076 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:35:00 crc kubenswrapper[4750]: E0309 19:35:00.376354 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:35:05 crc kubenswrapper[4750]: I0309 19:35:05.374024 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:35:05 crc kubenswrapper[4750]: E0309 19:35:05.374913 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:35:11 crc kubenswrapper[4750]: I0309 19:35:11.374252 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:35:11 crc kubenswrapper[4750]: E0309 19:35:11.375354 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:35:16 crc kubenswrapper[4750]: I0309 19:35:16.374100 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:35:17 crc kubenswrapper[4750]: I0309 19:35:17.603706 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6"} Mar 09 19:35:19 crc kubenswrapper[4750]: I0309 19:35:19.629763 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" exitCode=1 Mar 09 19:35:19 crc kubenswrapper[4750]: I0309 19:35:19.630108 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6"} Mar 09 19:35:19 crc kubenswrapper[4750]: I0309 19:35:19.630725 4750 scope.go:117] "RemoveContainer" containerID="8e7922b6b5245b80ecc12d83f76a77b155c3ca93e9cbef64b8e1d4cba1b3c205" Mar 09 19:35:19 crc kubenswrapper[4750]: I0309 19:35:19.631456 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:35:19 crc kubenswrapper[4750]: E0309 19:35:19.631718 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:35:20 crc kubenswrapper[4750]: I0309 19:35:20.189531 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:35:20 crc kubenswrapper[4750]: I0309 19:35:20.189898 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:35:20 crc kubenswrapper[4750]: I0309 19:35:20.189919 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:35:20 crc kubenswrapper[4750]: I0309 19:35:20.648799 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:35:20 crc kubenswrapper[4750]: E0309 19:35:20.649587 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:35:26 crc kubenswrapper[4750]: I0309 19:35:26.373116 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:35:26 crc kubenswrapper[4750]: I0309 19:35:26.720456 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"57fbac10ed6dc05021cd8f937ebb13f435963b2c8376f16b766d7a9c5fd75504"} Mar 09 19:35:32 crc kubenswrapper[4750]: I0309 19:35:32.373079 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:35:32 crc kubenswrapper[4750]: E0309 19:35:32.373872 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.837299 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z5vxm"] Mar 09 19:35:39 crc kubenswrapper[4750]: E0309 19:35:39.838455 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c511856-8984-4c7c-8af1-3c093025dce8" containerName="oc" Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.838469 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c511856-8984-4c7c-8af1-3c093025dce8" containerName="oc" Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.838729 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c511856-8984-4c7c-8af1-3c093025dce8" containerName="oc" Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.840073 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.850147 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5vxm"] Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.972473 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d497ab4a-bba2-411c-9ffc-7197fc96b0db-catalog-content\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.972530 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d497ab4a-bba2-411c-9ffc-7197fc96b0db-utilities\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:39 crc kubenswrapper[4750]: I0309 19:35:39.972702 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpvb\" (UniqueName: \"kubernetes.io/projected/d497ab4a-bba2-411c-9ffc-7197fc96b0db-kube-api-access-xmpvb\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.074193 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpvb\" (UniqueName: \"kubernetes.io/projected/d497ab4a-bba2-411c-9ffc-7197fc96b0db-kube-api-access-xmpvb\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.074318 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d497ab4a-bba2-411c-9ffc-7197fc96b0db-catalog-content\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.074354 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d497ab4a-bba2-411c-9ffc-7197fc96b0db-utilities\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.075324 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d497ab4a-bba2-411c-9ffc-7197fc96b0db-utilities\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.075400 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d497ab4a-bba2-411c-9ffc-7197fc96b0db-catalog-content\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.101962 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpvb\" (UniqueName: \"kubernetes.io/projected/d497ab4a-bba2-411c-9ffc-7197fc96b0db-kube-api-access-xmpvb\") pod \"certified-operators-z5vxm\" (UID: \"d497ab4a-bba2-411c-9ffc-7197fc96b0db\") " pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.169234 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.664202 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5vxm"] Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.880684 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5vxm" event={"ID":"d497ab4a-bba2-411c-9ffc-7197fc96b0db","Type":"ContainerStarted","Data":"ca2d538103aaa9196d826b2f526b60b23365c1a56625799fcedd87b3593436e1"} Mar 09 19:35:40 crc kubenswrapper[4750]: I0309 19:35:40.882113 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5vxm" event={"ID":"d497ab4a-bba2-411c-9ffc-7197fc96b0db","Type":"ContainerStarted","Data":"9d2ade01775f2c8296f8c975778f3e6a2d268009806ef0d077a93ca2a81b337c"} Mar 09 19:35:41 crc kubenswrapper[4750]: E0309 19:35:41.030653 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd497ab4a_bba2_411c_9ffc_7197fc96b0db.slice/crio-conmon-ca2d538103aaa9196d826b2f526b60b23365c1a56625799fcedd87b3593436e1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd497ab4a_bba2_411c_9ffc_7197fc96b0db.slice/crio-ca2d538103aaa9196d826b2f526b60b23365c1a56625799fcedd87b3593436e1.scope\": RecentStats: unable to find data in memory cache]" Mar 09 19:35:41 crc kubenswrapper[4750]: I0309 19:35:41.903232 4750 generic.go:334] "Generic (PLEG): container finished" podID="d497ab4a-bba2-411c-9ffc-7197fc96b0db" containerID="ca2d538103aaa9196d826b2f526b60b23365c1a56625799fcedd87b3593436e1" exitCode=0 Mar 09 19:35:41 crc kubenswrapper[4750]: I0309 19:35:41.903301 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5vxm" event={"ID":"d497ab4a-bba2-411c-9ffc-7197fc96b0db","Type":"ContainerDied","Data":"ca2d538103aaa9196d826b2f526b60b23365c1a56625799fcedd87b3593436e1"} Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.242189 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j5dds"] Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.246829 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.277154 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j5dds"] Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.329173 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-utilities\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.329335 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-catalog-content\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.329370 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9xjf\" (UniqueName: \"kubernetes.io/projected/9794f3f7-7bad-420b-b156-878ab1876b5d-kube-api-access-x9xjf\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.431664 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-utilities\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.431781 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-catalog-content\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.431810 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9xjf\" (UniqueName: \"kubernetes.io/projected/9794f3f7-7bad-420b-b156-878ab1876b5d-kube-api-access-x9xjf\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.432435 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-utilities\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.432483 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-catalog-content\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.464266 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9xjf\" (UniqueName: \"kubernetes.io/projected/9794f3f7-7bad-420b-b156-878ab1876b5d-kube-api-access-x9xjf\") pod \"redhat-marketplace-j5dds\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:42 crc kubenswrapper[4750]: I0309 19:35:42.573566 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:43 crc kubenswrapper[4750]: I0309 19:35:43.122509 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j5dds"] Mar 09 19:35:43 crc kubenswrapper[4750]: I0309 19:35:43.935724 4750 generic.go:334] "Generic (PLEG): container finished" podID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerID="1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a" exitCode=0 Mar 09 19:35:43 crc kubenswrapper[4750]: I0309 19:35:43.935948 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j5dds" event={"ID":"9794f3f7-7bad-420b-b156-878ab1876b5d","Type":"ContainerDied","Data":"1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a"} Mar 09 19:35:43 crc kubenswrapper[4750]: I0309 19:35:43.936354 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j5dds" event={"ID":"9794f3f7-7bad-420b-b156-878ab1876b5d","Type":"ContainerStarted","Data":"84719ba9b35215e88420cac1a365c7a9284fdf66e4fa4144d5d66befd2a72c69"} Mar 09 19:35:44 crc kubenswrapper[4750]: I0309 19:35:44.950120 4750 generic.go:334] "Generic (PLEG): container finished" podID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerID="fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24" exitCode=0 Mar 09 19:35:44 crc kubenswrapper[4750]: I0309 19:35:44.950155 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j5dds" event={"ID":"9794f3f7-7bad-420b-b156-878ab1876b5d","Type":"ContainerDied","Data":"fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24"} Mar 09 19:35:46 crc kubenswrapper[4750]: I0309 19:35:46.374365 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:35:46 crc kubenswrapper[4750]: E0309 19:35:46.374611 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:35:47 crc kubenswrapper[4750]: I0309 19:35:47.990932 4750 generic.go:334] "Generic (PLEG): container finished" podID="d497ab4a-bba2-411c-9ffc-7197fc96b0db" containerID="9bf11f9ff648178201e6f0d69b8309d2807440d9e51b73d83ec4ba907cd213fd" exitCode=0 Mar 09 19:35:47 crc kubenswrapper[4750]: I0309 19:35:47.991019 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5vxm" event={"ID":"d497ab4a-bba2-411c-9ffc-7197fc96b0db","Type":"ContainerDied","Data":"9bf11f9ff648178201e6f0d69b8309d2807440d9e51b73d83ec4ba907cd213fd"} Mar 09 19:35:47 crc kubenswrapper[4750]: I0309 19:35:47.997590 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j5dds" event={"ID":"9794f3f7-7bad-420b-b156-878ab1876b5d","Type":"ContainerStarted","Data":"a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231"} Mar 09 19:35:48 crc kubenswrapper[4750]: I0309 19:35:48.059388 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j5dds" podStartSLOduration=2.791973615 podStartE2EDuration="6.059362123s" podCreationTimestamp="2026-03-09 19:35:42 +0000 UTC" firstStartedPulling="2026-03-09 19:35:43.942953055 +0000 UTC m=+4225.285425453" lastFinishedPulling="2026-03-09 19:35:47.210341563 +0000 UTC m=+4228.552813961" observedRunningTime="2026-03-09 19:35:48.043380652 +0000 UTC m=+4229.385853070" watchObservedRunningTime="2026-03-09 19:35:48.059362123 +0000 UTC m=+4229.401834541" Mar 09 19:35:49 crc kubenswrapper[4750]: I0309 19:35:49.011002 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5vxm" event={"ID":"d497ab4a-bba2-411c-9ffc-7197fc96b0db","Type":"ContainerStarted","Data":"56e9758f1070938d99633fec464d762432f0b7dd8ee29eb712ca672444a6fd19"} Mar 09 19:35:49 crc kubenswrapper[4750]: I0309 19:35:49.038737 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z5vxm" podStartSLOduration=3.547206711 podStartE2EDuration="10.03871831s" podCreationTimestamp="2026-03-09 19:35:39 +0000 UTC" firstStartedPulling="2026-03-09 19:35:41.911914829 +0000 UTC m=+4223.254387267" lastFinishedPulling="2026-03-09 19:35:48.403426428 +0000 UTC m=+4229.745898866" observedRunningTime="2026-03-09 19:35:49.032495742 +0000 UTC m=+4230.374968150" watchObservedRunningTime="2026-03-09 19:35:49.03871831 +0000 UTC m=+4230.381190718" Mar 09 19:35:50 crc kubenswrapper[4750]: I0309 19:35:50.170073 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:50 crc kubenswrapper[4750]: I0309 19:35:50.170326 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:35:51 crc kubenswrapper[4750]: I0309 19:35:51.521456 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-z5vxm" podUID="d497ab4a-bba2-411c-9ffc-7197fc96b0db" containerName="registry-server" probeResult="failure" output=< Mar 09 19:35:51 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:35:51 crc kubenswrapper[4750]: > Mar 09 19:35:52 crc kubenswrapper[4750]: I0309 19:35:52.573728 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:52 crc kubenswrapper[4750]: I0309 19:35:52.574120 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:52 crc kubenswrapper[4750]: I0309 19:35:52.637799 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:53 crc kubenswrapper[4750]: I0309 19:35:53.152525 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:54 crc kubenswrapper[4750]: I0309 19:35:54.825006 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j5dds"] Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.100662 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j5dds" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="registry-server" containerID="cri-o://a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231" gracePeriod=2 Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.692711 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.803895 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9xjf\" (UniqueName: \"kubernetes.io/projected/9794f3f7-7bad-420b-b156-878ab1876b5d-kube-api-access-x9xjf\") pod \"9794f3f7-7bad-420b-b156-878ab1876b5d\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.803987 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-utilities\") pod \"9794f3f7-7bad-420b-b156-878ab1876b5d\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.804190 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-catalog-content\") pod \"9794f3f7-7bad-420b-b156-878ab1876b5d\" (UID: \"9794f3f7-7bad-420b-b156-878ab1876b5d\") " Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.805132 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-utilities" (OuterVolumeSpecName: "utilities") pod "9794f3f7-7bad-420b-b156-878ab1876b5d" (UID: "9794f3f7-7bad-420b-b156-878ab1876b5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.813813 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9794f3f7-7bad-420b-b156-878ab1876b5d-kube-api-access-x9xjf" (OuterVolumeSpecName: "kube-api-access-x9xjf") pod "9794f3f7-7bad-420b-b156-878ab1876b5d" (UID: "9794f3f7-7bad-420b-b156-878ab1876b5d"). InnerVolumeSpecName "kube-api-access-x9xjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.840768 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9794f3f7-7bad-420b-b156-878ab1876b5d" (UID: "9794f3f7-7bad-420b-b156-878ab1876b5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.906792 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9xjf\" (UniqueName: \"kubernetes.io/projected/9794f3f7-7bad-420b-b156-878ab1876b5d-kube-api-access-x9xjf\") on node \"crc\" DevicePath \"\"" Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.906827 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:35:56 crc kubenswrapper[4750]: I0309 19:35:56.906837 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9794f3f7-7bad-420b-b156-878ab1876b5d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.121897 4750 generic.go:334] "Generic (PLEG): container finished" podID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerID="a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231" exitCode=0 Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.121960 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j5dds" event={"ID":"9794f3f7-7bad-420b-b156-878ab1876b5d","Type":"ContainerDied","Data":"a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231"} Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.121995 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j5dds" event={"ID":"9794f3f7-7bad-420b-b156-878ab1876b5d","Type":"ContainerDied","Data":"84719ba9b35215e88420cac1a365c7a9284fdf66e4fa4144d5d66befd2a72c69"} Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.122010 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j5dds" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.122016 4750 scope.go:117] "RemoveContainer" containerID="a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.177950 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j5dds"] Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.188404 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j5dds"] Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.190356 4750 scope.go:117] "RemoveContainer" containerID="fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.227794 4750 scope.go:117] "RemoveContainer" containerID="1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.278351 4750 scope.go:117] "RemoveContainer" containerID="a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231" Mar 09 19:35:57 crc kubenswrapper[4750]: E0309 19:35:57.279604 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231\": container with ID starting with a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231 not found: ID does not exist" containerID="a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.279688 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231"} err="failed to get container status \"a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231\": rpc error: code = NotFound desc = could not find container \"a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231\": container with ID starting with a138bd97ed359a881d6e7f5765dced42e68301d6178c6ba16bc34700b2618231 not found: ID does not exist" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.279739 4750 scope.go:117] "RemoveContainer" containerID="fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24" Mar 09 19:35:57 crc kubenswrapper[4750]: E0309 19:35:57.280177 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24\": container with ID starting with fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24 not found: ID does not exist" containerID="fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.280235 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24"} err="failed to get container status \"fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24\": rpc error: code = NotFound desc = could not find container \"fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24\": container with ID starting with fc25cd51cb75e1f746ef7b183f4c0b37a9a0c1e2f889521ebec907b68b61be24 not found: ID does not exist" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.280269 4750 scope.go:117] "RemoveContainer" containerID="1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a" Mar 09 19:35:57 crc kubenswrapper[4750]: E0309 19:35:57.280602 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a\": container with ID starting with 1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a not found: ID does not exist" containerID="1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.280687 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a"} err="failed to get container status \"1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a\": rpc error: code = NotFound desc = could not find container \"1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a\": container with ID starting with 1a4c2a8a79da259b67011d54b5f2e5f8bf6d04c0b2c9b1a19768f0bbba92f28a not found: ID does not exist" Mar 09 19:35:57 crc kubenswrapper[4750]: I0309 19:35:57.404052 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" path="/var/lib/kubelet/pods/9794f3f7-7bad-420b-b156-878ab1876b5d/volumes" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.166238 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551416-8w8ff"] Mar 09 19:36:00 crc kubenswrapper[4750]: E0309 19:36:00.166967 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="registry-server" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.166991 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="registry-server" Mar 09 19:36:00 crc kubenswrapper[4750]: E0309 19:36:00.167015 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="extract-content" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.167028 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="extract-content" Mar 09 19:36:00 crc kubenswrapper[4750]: E0309 19:36:00.167053 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="extract-utilities" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.167065 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="extract-utilities" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.167452 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9794f3f7-7bad-420b-b156-878ab1876b5d" containerName="registry-server" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.169348 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551416-8w8ff" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.172439 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.172443 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.175195 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.188122 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551416-8w8ff"] Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.257596 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.285309 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9rjq\" (UniqueName: \"kubernetes.io/projected/93ad7de8-ab8a-4f75-94fd-112d78659dc6-kube-api-access-q9rjq\") pod \"auto-csr-approver-29551416-8w8ff\" (UID: \"93ad7de8-ab8a-4f75-94fd-112d78659dc6\") " pod="openshift-infra/auto-csr-approver-29551416-8w8ff" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.333427 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z5vxm" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.373799 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:36:00 crc kubenswrapper[4750]: E0309 19:36:00.374066 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.387928 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9rjq\" (UniqueName: \"kubernetes.io/projected/93ad7de8-ab8a-4f75-94fd-112d78659dc6-kube-api-access-q9rjq\") pod \"auto-csr-approver-29551416-8w8ff\" (UID: \"93ad7de8-ab8a-4f75-94fd-112d78659dc6\") " pod="openshift-infra/auto-csr-approver-29551416-8w8ff" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.412820 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9rjq\" (UniqueName: \"kubernetes.io/projected/93ad7de8-ab8a-4f75-94fd-112d78659dc6-kube-api-access-q9rjq\") pod \"auto-csr-approver-29551416-8w8ff\" (UID: \"93ad7de8-ab8a-4f75-94fd-112d78659dc6\") " pod="openshift-infra/auto-csr-approver-29551416-8w8ff" Mar 09 19:36:00 crc kubenswrapper[4750]: I0309 19:36:00.517186 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551416-8w8ff" Mar 09 19:36:01 crc kubenswrapper[4750]: I0309 19:36:01.019805 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551416-8w8ff"] Mar 09 19:36:01 crc kubenswrapper[4750]: W0309 19:36:01.031850 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ad7de8_ab8a_4f75_94fd_112d78659dc6.slice/crio-357f125fcd928297044335549f4a74e8501c771187a4fab8e5b5933264d29528 WatchSource:0}: Error finding container 357f125fcd928297044335549f4a74e8501c771187a4fab8e5b5933264d29528: Status 404 returned error can't find the container with id 357f125fcd928297044335549f4a74e8501c771187a4fab8e5b5933264d29528 Mar 09 19:36:01 crc kubenswrapper[4750]: I0309 19:36:01.054142 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5vxm"] Mar 09 19:36:01 crc kubenswrapper[4750]: I0309 19:36:01.190147 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551416-8w8ff" event={"ID":"93ad7de8-ab8a-4f75-94fd-112d78659dc6","Type":"ContainerStarted","Data":"357f125fcd928297044335549f4a74e8501c771187a4fab8e5b5933264d29528"} Mar 09 19:36:01 crc kubenswrapper[4750]: I0309 19:36:01.223618 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rz4tq"] Mar 09 19:36:01 crc kubenswrapper[4750]: I0309 19:36:01.224016 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rz4tq" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="registry-server" containerID="cri-o://f613f84965c8c9ebc5ebaa0bc26191f9e1757e8e49807833c1c926f3c4fb9f31" gracePeriod=2 Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.231051 4750 generic.go:334] "Generic (PLEG): container finished" podID="c32787f9-abef-430d-8860-b27f40363bc9" containerID="f613f84965c8c9ebc5ebaa0bc26191f9e1757e8e49807833c1c926f3c4fb9f31" exitCode=0 Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.232528 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4tq" event={"ID":"c32787f9-abef-430d-8860-b27f40363bc9","Type":"ContainerDied","Data":"f613f84965c8c9ebc5ebaa0bc26191f9e1757e8e49807833c1c926f3c4fb9f31"} Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.232571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4tq" event={"ID":"c32787f9-abef-430d-8860-b27f40363bc9","Type":"ContainerDied","Data":"fc682976c7c5560704610d448ea67a33afa1cc93ecf1b04e36bb6e833fde0bc8"} Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.232588 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc682976c7c5560704610d448ea67a33afa1cc93ecf1b04e36bb6e833fde0bc8" Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.298111 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.447472 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2prf\" (UniqueName: \"kubernetes.io/projected/c32787f9-abef-430d-8860-b27f40363bc9-kube-api-access-z2prf\") pod \"c32787f9-abef-430d-8860-b27f40363bc9\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.447620 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-catalog-content\") pod \"c32787f9-abef-430d-8860-b27f40363bc9\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.447683 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-utilities\") pod \"c32787f9-abef-430d-8860-b27f40363bc9\" (UID: \"c32787f9-abef-430d-8860-b27f40363bc9\") " Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.448360 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-utilities" (OuterVolumeSpecName: "utilities") pod "c32787f9-abef-430d-8860-b27f40363bc9" (UID: "c32787f9-abef-430d-8860-b27f40363bc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.453378 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c32787f9-abef-430d-8860-b27f40363bc9-kube-api-access-z2prf" (OuterVolumeSpecName: "kube-api-access-z2prf") pod "c32787f9-abef-430d-8860-b27f40363bc9" (UID: "c32787f9-abef-430d-8860-b27f40363bc9"). InnerVolumeSpecName "kube-api-access-z2prf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.516149 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c32787f9-abef-430d-8860-b27f40363bc9" (UID: "c32787f9-abef-430d-8860-b27f40363bc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.549722 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2prf\" (UniqueName: \"kubernetes.io/projected/c32787f9-abef-430d-8860-b27f40363bc9-kube-api-access-z2prf\") on node \"crc\" DevicePath \"\"" Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.549750 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:36:02 crc kubenswrapper[4750]: I0309 19:36:02.549759 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32787f9-abef-430d-8860-b27f40363bc9-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:36:03 crc kubenswrapper[4750]: I0309 19:36:03.242844 4750 generic.go:334] "Generic (PLEG): container finished" podID="93ad7de8-ab8a-4f75-94fd-112d78659dc6" containerID="93443c9813a87180a82b2ac150ca191c0d9ec17584a05cf51066b162ba327618" exitCode=0 Mar 09 19:36:03 crc kubenswrapper[4750]: I0309 19:36:03.242923 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551416-8w8ff" event={"ID":"93ad7de8-ab8a-4f75-94fd-112d78659dc6","Type":"ContainerDied","Data":"93443c9813a87180a82b2ac150ca191c0d9ec17584a05cf51066b162ba327618"} Mar 09 19:36:03 crc kubenswrapper[4750]: I0309 19:36:03.242942 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4tq" Mar 09 19:36:03 crc kubenswrapper[4750]: I0309 19:36:03.294553 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rz4tq"] Mar 09 19:36:03 crc kubenswrapper[4750]: I0309 19:36:03.309018 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rz4tq"] Mar 09 19:36:03 crc kubenswrapper[4750]: I0309 19:36:03.384649 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c32787f9-abef-430d-8860-b27f40363bc9" path="/var/lib/kubelet/pods/c32787f9-abef-430d-8860-b27f40363bc9/volumes" Mar 09 19:36:04 crc kubenswrapper[4750]: I0309 19:36:04.922038 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551416-8w8ff" Mar 09 19:36:05 crc kubenswrapper[4750]: I0309 19:36:05.105154 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9rjq\" (UniqueName: \"kubernetes.io/projected/93ad7de8-ab8a-4f75-94fd-112d78659dc6-kube-api-access-q9rjq\") pod \"93ad7de8-ab8a-4f75-94fd-112d78659dc6\" (UID: \"93ad7de8-ab8a-4f75-94fd-112d78659dc6\") " Mar 09 19:36:05 crc kubenswrapper[4750]: I0309 19:36:05.114592 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93ad7de8-ab8a-4f75-94fd-112d78659dc6-kube-api-access-q9rjq" (OuterVolumeSpecName: "kube-api-access-q9rjq") pod "93ad7de8-ab8a-4f75-94fd-112d78659dc6" (UID: "93ad7de8-ab8a-4f75-94fd-112d78659dc6"). InnerVolumeSpecName "kube-api-access-q9rjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:36:05 crc kubenswrapper[4750]: I0309 19:36:05.208213 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9rjq\" (UniqueName: \"kubernetes.io/projected/93ad7de8-ab8a-4f75-94fd-112d78659dc6-kube-api-access-q9rjq\") on node \"crc\" DevicePath \"\"" Mar 09 19:36:05 crc kubenswrapper[4750]: I0309 19:36:05.268131 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551416-8w8ff" event={"ID":"93ad7de8-ab8a-4f75-94fd-112d78659dc6","Type":"ContainerDied","Data":"357f125fcd928297044335549f4a74e8501c771187a4fab8e5b5933264d29528"} Mar 09 19:36:05 crc kubenswrapper[4750]: I0309 19:36:05.268193 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="357f125fcd928297044335549f4a74e8501c771187a4fab8e5b5933264d29528" Mar 09 19:36:05 crc kubenswrapper[4750]: I0309 19:36:05.268303 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551416-8w8ff" Mar 09 19:36:06 crc kubenswrapper[4750]: I0309 19:36:06.031284 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551410-5jvnp"] Mar 09 19:36:06 crc kubenswrapper[4750]: I0309 19:36:06.042228 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551410-5jvnp"] Mar 09 19:36:07 crc kubenswrapper[4750]: I0309 19:36:07.389768 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcf1b77f-8234-40b8-90be-b073864b5514" path="/var/lib/kubelet/pods/bcf1b77f-8234-40b8-90be-b073864b5514/volumes" Mar 09 19:36:11 crc kubenswrapper[4750]: I0309 19:36:11.374839 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:36:11 crc kubenswrapper[4750]: E0309 19:36:11.376365 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:36:22 crc kubenswrapper[4750]: I0309 19:36:22.373556 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:36:22 crc kubenswrapper[4750]: E0309 19:36:22.374240 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:36:33 crc kubenswrapper[4750]: I0309 19:36:33.375159 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:36:33 crc kubenswrapper[4750]: E0309 19:36:33.376334 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:36:47 crc kubenswrapper[4750]: I0309 19:36:47.374890 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:36:47 crc kubenswrapper[4750]: E0309 19:36:47.376383 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:36:47 crc kubenswrapper[4750]: I0309 19:36:47.772389 4750 scope.go:117] "RemoveContainer" containerID="7b5c76dd57685b3c8123c73c8585cc70b1702090f4fd1a86012fd80c0116017d" Mar 09 19:36:47 crc kubenswrapper[4750]: I0309 19:36:47.807697 4750 scope.go:117] "RemoveContainer" containerID="0194c2b57a19b4324297353a92635ad6352cc5df89591df85f62ca169ef773b1" Mar 09 19:36:47 crc kubenswrapper[4750]: I0309 19:36:47.898568 4750 scope.go:117] "RemoveContainer" containerID="f613f84965c8c9ebc5ebaa0bc26191f9e1757e8e49807833c1c926f3c4fb9f31" Mar 09 19:36:47 crc kubenswrapper[4750]: I0309 19:36:47.957023 4750 scope.go:117] "RemoveContainer" containerID="40d46426df3ceea5a8bd1803aad6998d54174a2a8f861dd3f6539be030b2443d" Mar 09 19:36:58 crc kubenswrapper[4750]: I0309 19:36:58.373704 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:36:58 crc kubenswrapper[4750]: E0309 19:36:58.374561 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:37:12 crc kubenswrapper[4750]: I0309 19:37:12.375255 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:37:12 crc kubenswrapper[4750]: E0309 19:37:12.376756 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:37:26 crc kubenswrapper[4750]: I0309 19:37:26.374882 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:37:26 crc kubenswrapper[4750]: E0309 19:37:26.376184 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:37:40 crc kubenswrapper[4750]: I0309 19:37:40.374225 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:37:40 crc kubenswrapper[4750]: E0309 19:37:40.375366 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:37:51 crc kubenswrapper[4750]: I0309 19:37:51.744340 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:37:51 crc kubenswrapper[4750]: I0309 19:37:51.745075 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:37:53 crc kubenswrapper[4750]: I0309 19:37:53.374974 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:37:53 crc kubenswrapper[4750]: E0309 19:37:53.375791 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.188100 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551418-9l4wb"] Mar 09 19:38:00 crc kubenswrapper[4750]: E0309 19:38:00.190906 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="registry-server" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.190928 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="registry-server" Mar 09 19:38:00 crc kubenswrapper[4750]: E0309 19:38:00.190953 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="extract-utilities" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.190967 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="extract-utilities" Mar 09 19:38:00 crc kubenswrapper[4750]: E0309 19:38:00.191000 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ad7de8-ab8a-4f75-94fd-112d78659dc6" containerName="oc" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.191009 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ad7de8-ab8a-4f75-94fd-112d78659dc6" containerName="oc" Mar 09 19:38:00 crc kubenswrapper[4750]: E0309 19:38:00.191038 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="extract-content" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.191049 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="extract-content" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.192157 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ad7de8-ab8a-4f75-94fd-112d78659dc6" containerName="oc" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.192301 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32787f9-abef-430d-8860-b27f40363bc9" containerName="registry-server" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.193525 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551418-9l4wb" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.196988 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.197506 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.202850 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.209140 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551418-9l4wb"] Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.349952 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vbtb\" (UniqueName: \"kubernetes.io/projected/675d07ed-482d-4b1e-b5ab-2553e5a64214-kube-api-access-6vbtb\") pod \"auto-csr-approver-29551418-9l4wb\" (UID: \"675d07ed-482d-4b1e-b5ab-2553e5a64214\") " pod="openshift-infra/auto-csr-approver-29551418-9l4wb" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.452006 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vbtb\" (UniqueName: \"kubernetes.io/projected/675d07ed-482d-4b1e-b5ab-2553e5a64214-kube-api-access-6vbtb\") pod \"auto-csr-approver-29551418-9l4wb\" (UID: \"675d07ed-482d-4b1e-b5ab-2553e5a64214\") " pod="openshift-infra/auto-csr-approver-29551418-9l4wb" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.476695 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vbtb\" (UniqueName: \"kubernetes.io/projected/675d07ed-482d-4b1e-b5ab-2553e5a64214-kube-api-access-6vbtb\") pod \"auto-csr-approver-29551418-9l4wb\" (UID: \"675d07ed-482d-4b1e-b5ab-2553e5a64214\") " pod="openshift-infra/auto-csr-approver-29551418-9l4wb" Mar 09 19:38:00 crc kubenswrapper[4750]: I0309 19:38:00.528897 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551418-9l4wb" Mar 09 19:38:01 crc kubenswrapper[4750]: I0309 19:38:01.062878 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551418-9l4wb"] Mar 09 19:38:01 crc kubenswrapper[4750]: I0309 19:38:01.065618 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:38:01 crc kubenswrapper[4750]: I0309 19:38:01.513943 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551418-9l4wb" event={"ID":"675d07ed-482d-4b1e-b5ab-2553e5a64214","Type":"ContainerStarted","Data":"6cd6a79c524b480b17cbdbe716906779b276cbccfebcbb3c8e82aa20818186d1"} Mar 09 19:38:03 crc kubenswrapper[4750]: I0309 19:38:03.538910 4750 generic.go:334] "Generic (PLEG): container finished" podID="675d07ed-482d-4b1e-b5ab-2553e5a64214" containerID="b1aa1aa93d21843cd6efea4bc6c5faab0b92a8c0f5b2b4452c0350f66c918542" exitCode=0 Mar 09 19:38:03 crc kubenswrapper[4750]: I0309 19:38:03.538980 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551418-9l4wb" event={"ID":"675d07ed-482d-4b1e-b5ab-2553e5a64214","Type":"ContainerDied","Data":"b1aa1aa93d21843cd6efea4bc6c5faab0b92a8c0f5b2b4452c0350f66c918542"} Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.028575 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551418-9l4wb" Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.110886 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vbtb\" (UniqueName: \"kubernetes.io/projected/675d07ed-482d-4b1e-b5ab-2553e5a64214-kube-api-access-6vbtb\") pod \"675d07ed-482d-4b1e-b5ab-2553e5a64214\" (UID: \"675d07ed-482d-4b1e-b5ab-2553e5a64214\") " Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.122985 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675d07ed-482d-4b1e-b5ab-2553e5a64214-kube-api-access-6vbtb" (OuterVolumeSpecName: "kube-api-access-6vbtb") pod "675d07ed-482d-4b1e-b5ab-2553e5a64214" (UID: "675d07ed-482d-4b1e-b5ab-2553e5a64214"). InnerVolumeSpecName "kube-api-access-6vbtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.213954 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vbtb\" (UniqueName: \"kubernetes.io/projected/675d07ed-482d-4b1e-b5ab-2553e5a64214-kube-api-access-6vbtb\") on node \"crc\" DevicePath \"\"" Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.373712 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:38:05 crc kubenswrapper[4750]: E0309 19:38:05.374084 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.565855 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551418-9l4wb" event={"ID":"675d07ed-482d-4b1e-b5ab-2553e5a64214","Type":"ContainerDied","Data":"6cd6a79c524b480b17cbdbe716906779b276cbccfebcbb3c8e82aa20818186d1"} Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.565951 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cd6a79c524b480b17cbdbe716906779b276cbccfebcbb3c8e82aa20818186d1" Mar 09 19:38:05 crc kubenswrapper[4750]: I0309 19:38:05.565945 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551418-9l4wb" Mar 09 19:38:06 crc kubenswrapper[4750]: I0309 19:38:06.131279 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551412-4qkjs"] Mar 09 19:38:06 crc kubenswrapper[4750]: I0309 19:38:06.144399 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551412-4qkjs"] Mar 09 19:38:07 crc kubenswrapper[4750]: I0309 19:38:07.385180 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5cdd0e0-e091-4a05-bab5-872aeef5d293" path="/var/lib/kubelet/pods/c5cdd0e0-e091-4a05-bab5-872aeef5d293/volumes" Mar 09 19:38:17 crc kubenswrapper[4750]: I0309 19:38:17.373141 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:38:17 crc kubenswrapper[4750]: E0309 19:38:17.373951 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:38:21 crc kubenswrapper[4750]: I0309 19:38:21.744324 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:38:21 crc kubenswrapper[4750]: I0309 19:38:21.745660 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:38:30 crc kubenswrapper[4750]: I0309 19:38:30.375805 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:38:30 crc kubenswrapper[4750]: E0309 19:38:30.376394 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:38:43 crc kubenswrapper[4750]: I0309 19:38:43.374235 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:38:43 crc kubenswrapper[4750]: E0309 19:38:43.375426 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:38:48 crc kubenswrapper[4750]: I0309 19:38:48.091600 4750 scope.go:117] "RemoveContainer" containerID="7d5be83a6dfd3a19ced9532ac56dc14e399874126bc63a039029797312850cb7" Mar 09 19:38:51 crc kubenswrapper[4750]: I0309 19:38:51.743745 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:38:51 crc kubenswrapper[4750]: I0309 19:38:51.744723 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:38:51 crc kubenswrapper[4750]: I0309 19:38:51.744819 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:38:51 crc kubenswrapper[4750]: I0309 19:38:51.747382 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"57fbac10ed6dc05021cd8f937ebb13f435963b2c8376f16b766d7a9c5fd75504"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:38:51 crc kubenswrapper[4750]: I0309 19:38:51.747859 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://57fbac10ed6dc05021cd8f937ebb13f435963b2c8376f16b766d7a9c5fd75504" gracePeriod=600 Mar 09 19:38:52 crc kubenswrapper[4750]: I0309 19:38:52.156839 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="57fbac10ed6dc05021cd8f937ebb13f435963b2c8376f16b766d7a9c5fd75504" exitCode=0 Mar 09 19:38:52 crc kubenswrapper[4750]: I0309 19:38:52.156887 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"57fbac10ed6dc05021cd8f937ebb13f435963b2c8376f16b766d7a9c5fd75504"} Mar 09 19:38:52 crc kubenswrapper[4750]: I0309 19:38:52.156958 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc"} Mar 09 19:38:52 crc kubenswrapper[4750]: I0309 19:38:52.156988 4750 scope.go:117] "RemoveContainer" containerID="cafcc084b68e38d647da62d2fd962f3c53fd9a965dc9d7ab3f595b38c068e9ca" Mar 09 19:38:57 crc kubenswrapper[4750]: I0309 19:38:57.373930 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:38:57 crc kubenswrapper[4750]: E0309 19:38:57.375375 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:39:08 crc kubenswrapper[4750]: I0309 19:39:08.374178 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:39:08 crc kubenswrapper[4750]: E0309 19:39:08.377067 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:39:19 crc kubenswrapper[4750]: I0309 19:39:19.387276 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:39:19 crc kubenswrapper[4750]: E0309 19:39:19.388145 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:39:30 crc kubenswrapper[4750]: I0309 19:39:30.373779 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:39:30 crc kubenswrapper[4750]: E0309 19:39:30.374977 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:39:41 crc kubenswrapper[4750]: I0309 19:39:41.373608 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:39:41 crc kubenswrapper[4750]: E0309 19:39:41.374479 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:39:56 crc kubenswrapper[4750]: I0309 19:39:56.374069 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:39:56 crc kubenswrapper[4750]: E0309 19:39:56.375213 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.177855 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551420-mj8jl"] Mar 09 19:40:00 crc kubenswrapper[4750]: E0309 19:40:00.179541 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675d07ed-482d-4b1e-b5ab-2553e5a64214" containerName="oc" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.179803 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="675d07ed-482d-4b1e-b5ab-2553e5a64214" containerName="oc" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.180362 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="675d07ed-482d-4b1e-b5ab-2553e5a64214" containerName="oc" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.182052 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551420-mj8jl" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.185139 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.186509 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.188410 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551420-mj8jl"] Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.190522 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.318884 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbxtm\" (UniqueName: \"kubernetes.io/projected/4207172d-d888-4a93-bba8-9316e3425d5b-kube-api-access-cbxtm\") pod \"auto-csr-approver-29551420-mj8jl\" (UID: \"4207172d-d888-4a93-bba8-9316e3425d5b\") " pod="openshift-infra/auto-csr-approver-29551420-mj8jl" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.421345 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbxtm\" (UniqueName: \"kubernetes.io/projected/4207172d-d888-4a93-bba8-9316e3425d5b-kube-api-access-cbxtm\") pod \"auto-csr-approver-29551420-mj8jl\" (UID: \"4207172d-d888-4a93-bba8-9316e3425d5b\") " pod="openshift-infra/auto-csr-approver-29551420-mj8jl" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.450767 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbxtm\" (UniqueName: \"kubernetes.io/projected/4207172d-d888-4a93-bba8-9316e3425d5b-kube-api-access-cbxtm\") pod \"auto-csr-approver-29551420-mj8jl\" (UID: \"4207172d-d888-4a93-bba8-9316e3425d5b\") " pod="openshift-infra/auto-csr-approver-29551420-mj8jl" Mar 09 19:40:00 crc kubenswrapper[4750]: I0309 19:40:00.509471 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551420-mj8jl" Mar 09 19:40:01 crc kubenswrapper[4750]: I0309 19:40:01.009378 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551420-mj8jl"] Mar 09 19:40:01 crc kubenswrapper[4750]: I0309 19:40:01.034164 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551420-mj8jl" event={"ID":"4207172d-d888-4a93-bba8-9316e3425d5b","Type":"ContainerStarted","Data":"c66fd8bf4d4dcd154f3d80146026a1dab9e97893870086884a5f11f0eec9cfcb"} Mar 09 19:40:03 crc kubenswrapper[4750]: I0309 19:40:03.064436 4750 generic.go:334] "Generic (PLEG): container finished" podID="4207172d-d888-4a93-bba8-9316e3425d5b" containerID="539cb87d4b09c1b2e472859625d392ec70d7f767301a563af43dcffa3ba1aae7" exitCode=0 Mar 09 19:40:03 crc kubenswrapper[4750]: I0309 19:40:03.064519 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551420-mj8jl" event={"ID":"4207172d-d888-4a93-bba8-9316e3425d5b","Type":"ContainerDied","Data":"539cb87d4b09c1b2e472859625d392ec70d7f767301a563af43dcffa3ba1aae7"} Mar 09 19:40:04 crc kubenswrapper[4750]: I0309 19:40:04.469126 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551420-mj8jl" Mar 09 19:40:04 crc kubenswrapper[4750]: I0309 19:40:04.634024 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbxtm\" (UniqueName: \"kubernetes.io/projected/4207172d-d888-4a93-bba8-9316e3425d5b-kube-api-access-cbxtm\") pod \"4207172d-d888-4a93-bba8-9316e3425d5b\" (UID: \"4207172d-d888-4a93-bba8-9316e3425d5b\") " Mar 09 19:40:05 crc kubenswrapper[4750]: I0309 19:40:05.091930 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551420-mj8jl" event={"ID":"4207172d-d888-4a93-bba8-9316e3425d5b","Type":"ContainerDied","Data":"c66fd8bf4d4dcd154f3d80146026a1dab9e97893870086884a5f11f0eec9cfcb"} Mar 09 19:40:05 crc kubenswrapper[4750]: I0309 19:40:05.092002 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c66fd8bf4d4dcd154f3d80146026a1dab9e97893870086884a5f11f0eec9cfcb" Mar 09 19:40:05 crc kubenswrapper[4750]: I0309 19:40:05.092011 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551420-mj8jl" Mar 09 19:40:05 crc kubenswrapper[4750]: I0309 19:40:05.263540 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4207172d-d888-4a93-bba8-9316e3425d5b-kube-api-access-cbxtm" (OuterVolumeSpecName: "kube-api-access-cbxtm") pod "4207172d-d888-4a93-bba8-9316e3425d5b" (UID: "4207172d-d888-4a93-bba8-9316e3425d5b"). InnerVolumeSpecName "kube-api-access-cbxtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:40:05 crc kubenswrapper[4750]: I0309 19:40:05.351940 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbxtm\" (UniqueName: \"kubernetes.io/projected/4207172d-d888-4a93-bba8-9316e3425d5b-kube-api-access-cbxtm\") on node \"crc\" DevicePath \"\"" Mar 09 19:40:05 crc kubenswrapper[4750]: I0309 19:40:05.543618 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551414-qv9bj"] Mar 09 19:40:05 crc kubenswrapper[4750]: I0309 19:40:05.552096 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551414-qv9bj"] Mar 09 19:40:07 crc kubenswrapper[4750]: I0309 19:40:07.393012 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c511856-8984-4c7c-8af1-3c093025dce8" path="/var/lib/kubelet/pods/8c511856-8984-4c7c-8af1-3c093025dce8/volumes" Mar 09 19:40:09 crc kubenswrapper[4750]: I0309 19:40:09.379509 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:40:09 crc kubenswrapper[4750]: E0309 19:40:09.381041 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:40:22 crc kubenswrapper[4750]: I0309 19:40:22.374157 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:40:23 crc kubenswrapper[4750]: I0309 19:40:23.340047 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3"} Mar 09 19:40:25 crc kubenswrapper[4750]: I0309 19:40:25.189474 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:40:26 crc kubenswrapper[4750]: I0309 19:40:26.389442 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" exitCode=1 Mar 09 19:40:26 crc kubenswrapper[4750]: I0309 19:40:26.389560 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3"} Mar 09 19:40:26 crc kubenswrapper[4750]: I0309 19:40:26.389811 4750 scope.go:117] "RemoveContainer" containerID="252dffaee06ddf640bd270d524081b0b5e3f9af083faac43251cb47a22fe12e6" Mar 09 19:40:26 crc kubenswrapper[4750]: I0309 19:40:26.391082 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:40:26 crc kubenswrapper[4750]: E0309 19:40:26.391677 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:40:29 crc kubenswrapper[4750]: I0309 19:40:29.189819 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:40:29 crc kubenswrapper[4750]: I0309 19:40:29.195656 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:40:29 crc kubenswrapper[4750]: E0309 19:40:29.196015 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:40:30 crc kubenswrapper[4750]: I0309 19:40:30.190306 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:40:30 crc kubenswrapper[4750]: I0309 19:40:30.193281 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:40:30 crc kubenswrapper[4750]: E0309 19:40:30.193978 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:40:41 crc kubenswrapper[4750]: I0309 19:40:41.374442 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:40:41 crc kubenswrapper[4750]: E0309 19:40:41.375803 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:40:48 crc kubenswrapper[4750]: I0309 19:40:48.256569 4750 scope.go:117] "RemoveContainer" containerID="e3a8c85deb2996c38883afeddb20c941be784426a909a0d83385f4f8c00e1003" Mar 09 19:40:52 crc kubenswrapper[4750]: I0309 19:40:52.374802 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:40:52 crc kubenswrapper[4750]: E0309 19:40:52.375802 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:41:04 crc kubenswrapper[4750]: I0309 19:41:04.375303 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:41:04 crc kubenswrapper[4750]: E0309 19:41:04.376906 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:41:17 crc kubenswrapper[4750]: I0309 19:41:17.374209 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:41:17 crc kubenswrapper[4750]: E0309 19:41:17.375122 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:41:21 crc kubenswrapper[4750]: I0309 19:41:21.743573 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:41:21 crc kubenswrapper[4750]: I0309 19:41:21.745526 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:41:31 crc kubenswrapper[4750]: I0309 19:41:31.373777 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:41:31 crc kubenswrapper[4750]: E0309 19:41:31.374588 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:41:46 crc kubenswrapper[4750]: I0309 19:41:46.374135 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:41:46 crc kubenswrapper[4750]: E0309 19:41:46.375392 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:41:51 crc kubenswrapper[4750]: I0309 19:41:51.743147 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:41:51 crc kubenswrapper[4750]: I0309 19:41:51.743752 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.186108 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551422-vv8tp"] Mar 09 19:42:00 crc kubenswrapper[4750]: E0309 19:42:00.190843 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4207172d-d888-4a93-bba8-9316e3425d5b" containerName="oc" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.190887 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4207172d-d888-4a93-bba8-9316e3425d5b" containerName="oc" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.191374 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4207172d-d888-4a93-bba8-9316e3425d5b" containerName="oc" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.192383 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.194730 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.195402 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.197855 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.213221 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551422-vv8tp"] Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.283412 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcvwk\" (UniqueName: \"kubernetes.io/projected/e9c60766-4c58-41b5-b11e-5b3f22578968-kube-api-access-dcvwk\") pod \"auto-csr-approver-29551422-vv8tp\" (UID: \"e9c60766-4c58-41b5-b11e-5b3f22578968\") " pod="openshift-infra/auto-csr-approver-29551422-vv8tp" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.385867 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcvwk\" (UniqueName: \"kubernetes.io/projected/e9c60766-4c58-41b5-b11e-5b3f22578968-kube-api-access-dcvwk\") pod \"auto-csr-approver-29551422-vv8tp\" (UID: \"e9c60766-4c58-41b5-b11e-5b3f22578968\") " pod="openshift-infra/auto-csr-approver-29551422-vv8tp" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.416161 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcvwk\" (UniqueName: \"kubernetes.io/projected/e9c60766-4c58-41b5-b11e-5b3f22578968-kube-api-access-dcvwk\") pod \"auto-csr-approver-29551422-vv8tp\" (UID: \"e9c60766-4c58-41b5-b11e-5b3f22578968\") " pod="openshift-infra/auto-csr-approver-29551422-vv8tp" Mar 09 19:42:00 crc kubenswrapper[4750]: I0309 19:42:00.523457 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" Mar 09 19:42:01 crc kubenswrapper[4750]: I0309 19:42:01.058127 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551422-vv8tp"] Mar 09 19:42:01 crc kubenswrapper[4750]: I0309 19:42:01.374412 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:42:01 crc kubenswrapper[4750]: E0309 19:42:01.374944 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:42:01 crc kubenswrapper[4750]: I0309 19:42:01.555520 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" event={"ID":"e9c60766-4c58-41b5-b11e-5b3f22578968","Type":"ContainerStarted","Data":"553c554c3faebae06e896cea02ab837feb217a1c3bb94d7c99a07c5f7995ba1f"} Mar 09 19:42:02 crc kubenswrapper[4750]: I0309 19:42:02.568042 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" event={"ID":"e9c60766-4c58-41b5-b11e-5b3f22578968","Type":"ContainerStarted","Data":"a7a390f650ca358270660b6eea88462a92367e7537d06fb527600069366bb7dc"} Mar 09 19:42:02 crc kubenswrapper[4750]: I0309 19:42:02.595181 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" podStartSLOduration=1.671893311 podStartE2EDuration="2.595159974s" podCreationTimestamp="2026-03-09 19:42:00 +0000 UTC" firstStartedPulling="2026-03-09 19:42:01.062237486 +0000 UTC m=+4602.404709884" lastFinishedPulling="2026-03-09 19:42:01.985504149 +0000 UTC m=+4603.327976547" observedRunningTime="2026-03-09 19:42:02.588200745 +0000 UTC m=+4603.930673153" watchObservedRunningTime="2026-03-09 19:42:02.595159974 +0000 UTC m=+4603.937632382" Mar 09 19:42:03 crc kubenswrapper[4750]: I0309 19:42:03.583256 4750 generic.go:334] "Generic (PLEG): container finished" podID="e9c60766-4c58-41b5-b11e-5b3f22578968" containerID="a7a390f650ca358270660b6eea88462a92367e7537d06fb527600069366bb7dc" exitCode=0 Mar 09 19:42:03 crc kubenswrapper[4750]: I0309 19:42:03.583342 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" event={"ID":"e9c60766-4c58-41b5-b11e-5b3f22578968","Type":"ContainerDied","Data":"a7a390f650ca358270660b6eea88462a92367e7537d06fb527600069366bb7dc"} Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.015870 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.112443 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcvwk\" (UniqueName: \"kubernetes.io/projected/e9c60766-4c58-41b5-b11e-5b3f22578968-kube-api-access-dcvwk\") pod \"e9c60766-4c58-41b5-b11e-5b3f22578968\" (UID: \"e9c60766-4c58-41b5-b11e-5b3f22578968\") " Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.121829 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9c60766-4c58-41b5-b11e-5b3f22578968-kube-api-access-dcvwk" (OuterVolumeSpecName: "kube-api-access-dcvwk") pod "e9c60766-4c58-41b5-b11e-5b3f22578968" (UID: "e9c60766-4c58-41b5-b11e-5b3f22578968"). InnerVolumeSpecName "kube-api-access-dcvwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.215866 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcvwk\" (UniqueName: \"kubernetes.io/projected/e9c60766-4c58-41b5-b11e-5b3f22578968-kube-api-access-dcvwk\") on node \"crc\" DevicePath \"\"" Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.604839 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" event={"ID":"e9c60766-4c58-41b5-b11e-5b3f22578968","Type":"ContainerDied","Data":"553c554c3faebae06e896cea02ab837feb217a1c3bb94d7c99a07c5f7995ba1f"} Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.605081 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="553c554c3faebae06e896cea02ab837feb217a1c3bb94d7c99a07c5f7995ba1f" Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.604959 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551422-vv8tp" Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.695925 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551416-8w8ff"] Mar 09 19:42:05 crc kubenswrapper[4750]: I0309 19:42:05.703488 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551416-8w8ff"] Mar 09 19:42:07 crc kubenswrapper[4750]: I0309 19:42:07.387426 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93ad7de8-ab8a-4f75-94fd-112d78659dc6" path="/var/lib/kubelet/pods/93ad7de8-ab8a-4f75-94fd-112d78659dc6/volumes" Mar 09 19:42:16 crc kubenswrapper[4750]: I0309 19:42:16.377177 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:42:16 crc kubenswrapper[4750]: E0309 19:42:16.379747 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:42:21 crc kubenswrapper[4750]: I0309 19:42:21.743996 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:42:21 crc kubenswrapper[4750]: I0309 19:42:21.746854 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:42:21 crc kubenswrapper[4750]: I0309 19:42:21.747219 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:42:21 crc kubenswrapper[4750]: I0309 19:42:21.749042 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:42:21 crc kubenswrapper[4750]: I0309 19:42:21.749428 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" gracePeriod=600 Mar 09 19:42:21 crc kubenswrapper[4750]: E0309 19:42:21.884449 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:42:22 crc kubenswrapper[4750]: I0309 19:42:22.832814 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" exitCode=0 Mar 09 19:42:22 crc kubenswrapper[4750]: I0309 19:42:22.832908 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc"} Mar 09 19:42:22 crc kubenswrapper[4750]: I0309 19:42:22.833230 4750 scope.go:117] "RemoveContainer" containerID="57fbac10ed6dc05021cd8f937ebb13f435963b2c8376f16b766d7a9c5fd75504" Mar 09 19:42:22 crc kubenswrapper[4750]: I0309 19:42:22.833900 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:42:22 crc kubenswrapper[4750]: E0309 19:42:22.835145 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:42:28 crc kubenswrapper[4750]: I0309 19:42:28.374712 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:42:28 crc kubenswrapper[4750]: E0309 19:42:28.376229 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:42:38 crc kubenswrapper[4750]: I0309 19:42:38.373410 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:42:38 crc kubenswrapper[4750]: E0309 19:42:38.374215 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:42:39 crc kubenswrapper[4750]: I0309 19:42:39.381350 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:42:39 crc kubenswrapper[4750]: E0309 19:42:39.381940 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:42:48 crc kubenswrapper[4750]: I0309 19:42:48.414468 4750 scope.go:117] "RemoveContainer" containerID="93443c9813a87180a82b2ac150ca191c0d9ec17584a05cf51066b162ba327618" Mar 09 19:42:51 crc kubenswrapper[4750]: I0309 19:42:51.373599 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:42:51 crc kubenswrapper[4750]: E0309 19:42:51.374300 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:42:54 crc kubenswrapper[4750]: I0309 19:42:54.373380 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:42:54 crc kubenswrapper[4750]: E0309 19:42:54.374333 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:43:06 crc kubenswrapper[4750]: I0309 19:43:06.373975 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:43:06 crc kubenswrapper[4750]: I0309 19:43:06.374705 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:43:06 crc kubenswrapper[4750]: E0309 19:43:06.375084 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:43:06 crc kubenswrapper[4750]: E0309 19:43:06.375137 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:43:17 crc kubenswrapper[4750]: I0309 19:43:17.374056 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:43:17 crc kubenswrapper[4750]: E0309 19:43:17.374905 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:43:21 crc kubenswrapper[4750]: I0309 19:43:21.373953 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:43:21 crc kubenswrapper[4750]: E0309 19:43:21.375199 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:43:31 crc kubenswrapper[4750]: I0309 19:43:31.373525 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:43:31 crc kubenswrapper[4750]: E0309 19:43:31.374421 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:43:35 crc kubenswrapper[4750]: I0309 19:43:35.373767 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:43:35 crc kubenswrapper[4750]: E0309 19:43:35.374914 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.012083 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bhdx5"] Mar 09 19:43:39 crc kubenswrapper[4750]: E0309 19:43:39.013100 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c60766-4c58-41b5-b11e-5b3f22578968" containerName="oc" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.013131 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c60766-4c58-41b5-b11e-5b3f22578968" containerName="oc" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.013507 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9c60766-4c58-41b5-b11e-5b3f22578968" containerName="oc" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.016197 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.028172 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bhdx5"] Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.127938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtntl\" (UniqueName: \"kubernetes.io/projected/eb5bc9f7-34d2-4180-8633-5deac70271d8-kube-api-access-gtntl\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.128422 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5bc9f7-34d2-4180-8633-5deac70271d8-catalog-content\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.128610 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5bc9f7-34d2-4180-8633-5deac70271d8-utilities\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.231190 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtntl\" (UniqueName: \"kubernetes.io/projected/eb5bc9f7-34d2-4180-8633-5deac70271d8-kube-api-access-gtntl\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.231305 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5bc9f7-34d2-4180-8633-5deac70271d8-catalog-content\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.231392 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5bc9f7-34d2-4180-8633-5deac70271d8-utilities\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.232032 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5bc9f7-34d2-4180-8633-5deac70271d8-utilities\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.232162 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5bc9f7-34d2-4180-8633-5deac70271d8-catalog-content\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.253264 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtntl\" (UniqueName: \"kubernetes.io/projected/eb5bc9f7-34d2-4180-8633-5deac70271d8-kube-api-access-gtntl\") pod \"redhat-operators-bhdx5\" (UID: \"eb5bc9f7-34d2-4180-8633-5deac70271d8\") " pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.344336 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.835493 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bhdx5"] Mar 09 19:43:39 crc kubenswrapper[4750]: I0309 19:43:39.913287 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhdx5" event={"ID":"eb5bc9f7-34d2-4180-8633-5deac70271d8","Type":"ContainerStarted","Data":"d499f2addddcbfd869cb67f0af145da80c2dd6548a5a55c505d0588ec8727080"} Mar 09 19:43:40 crc kubenswrapper[4750]: I0309 19:43:40.926366 4750 generic.go:334] "Generic (PLEG): container finished" podID="eb5bc9f7-34d2-4180-8633-5deac70271d8" containerID="27e087d2cfb091e03cac282715ce7e209d09b2521426bfb46afbb60dd1ed7e5f" exitCode=0 Mar 09 19:43:40 crc kubenswrapper[4750]: I0309 19:43:40.926462 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhdx5" event={"ID":"eb5bc9f7-34d2-4180-8633-5deac70271d8","Type":"ContainerDied","Data":"27e087d2cfb091e03cac282715ce7e209d09b2521426bfb46afbb60dd1ed7e5f"} Mar 09 19:43:40 crc kubenswrapper[4750]: I0309 19:43:40.929381 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.391607 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rsmwj"] Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.394059 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.401355 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsmwj"] Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.500293 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-utilities\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.500332 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh9p5\" (UniqueName: \"kubernetes.io/projected/da237bf5-477b-4725-9388-55c35bcbfd30-kube-api-access-hh9p5\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.500498 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-catalog-content\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.602697 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-catalog-content\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.602867 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-utilities\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.602892 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh9p5\" (UniqueName: \"kubernetes.io/projected/da237bf5-477b-4725-9388-55c35bcbfd30-kube-api-access-hh9p5\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.603139 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-catalog-content\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.603448 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-utilities\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.621863 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh9p5\" (UniqueName: \"kubernetes.io/projected/da237bf5-477b-4725-9388-55c35bcbfd30-kube-api-access-hh9p5\") pod \"community-operators-rsmwj\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:42 crc kubenswrapper[4750]: I0309 19:43:42.720047 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:43:43 crc kubenswrapper[4750]: I0309 19:43:43.293414 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsmwj"] Mar 09 19:43:43 crc kubenswrapper[4750]: I0309 19:43:43.374329 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:43:43 crc kubenswrapper[4750]: E0309 19:43:43.374662 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:43:43 crc kubenswrapper[4750]: W0309 19:43:43.770020 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda237bf5_477b_4725_9388_55c35bcbfd30.slice/crio-8f2fa6cca48976af1e4d2bcb175f9c114cc0c0e0bf06f685b254a8baebc37536 WatchSource:0}: Error finding container 8f2fa6cca48976af1e4d2bcb175f9c114cc0c0e0bf06f685b254a8baebc37536: Status 404 returned error can't find the container with id 8f2fa6cca48976af1e4d2bcb175f9c114cc0c0e0bf06f685b254a8baebc37536 Mar 09 19:43:43 crc kubenswrapper[4750]: I0309 19:43:43.953822 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsmwj" event={"ID":"da237bf5-477b-4725-9388-55c35bcbfd30","Type":"ContainerStarted","Data":"8f2fa6cca48976af1e4d2bcb175f9c114cc0c0e0bf06f685b254a8baebc37536"} Mar 09 19:43:44 crc kubenswrapper[4750]: I0309 19:43:44.970440 4750 generic.go:334] "Generic (PLEG): container finished" podID="da237bf5-477b-4725-9388-55c35bcbfd30" containerID="a13d174f807bfd6825f31d04628faeacb781f227114544fa900803fe9beb17ee" exitCode=0 Mar 09 19:43:44 crc kubenswrapper[4750]: I0309 19:43:44.970523 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsmwj" event={"ID":"da237bf5-477b-4725-9388-55c35bcbfd30","Type":"ContainerDied","Data":"a13d174f807bfd6825f31d04628faeacb781f227114544fa900803fe9beb17ee"} Mar 09 19:43:45 crc kubenswrapper[4750]: I0309 19:43:45.984737 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsmwj" event={"ID":"da237bf5-477b-4725-9388-55c35bcbfd30","Type":"ContainerStarted","Data":"94c137630b437b4249dcc3bba6e61968b9a530aa015fa4e1c8a55196fa35961a"} Mar 09 19:43:48 crc kubenswrapper[4750]: I0309 19:43:48.019957 4750 generic.go:334] "Generic (PLEG): container finished" podID="da237bf5-477b-4725-9388-55c35bcbfd30" containerID="94c137630b437b4249dcc3bba6e61968b9a530aa015fa4e1c8a55196fa35961a" exitCode=0 Mar 09 19:43:48 crc kubenswrapper[4750]: I0309 19:43:48.020483 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsmwj" event={"ID":"da237bf5-477b-4725-9388-55c35bcbfd30","Type":"ContainerDied","Data":"94c137630b437b4249dcc3bba6e61968b9a530aa015fa4e1c8a55196fa35961a"} Mar 09 19:43:50 crc kubenswrapper[4750]: I0309 19:43:50.373253 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:43:50 crc kubenswrapper[4750]: E0309 19:43:50.373794 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:43:54 crc kubenswrapper[4750]: I0309 19:43:54.373763 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:43:54 crc kubenswrapper[4750]: E0309 19:43:54.374584 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:43:55 crc kubenswrapper[4750]: I0309 19:43:55.098022 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsmwj" event={"ID":"da237bf5-477b-4725-9388-55c35bcbfd30","Type":"ContainerStarted","Data":"e79cc5318b76c037202f6c41fc0f1eba87af9af5ac3743eccc42e8d9e555451a"} Mar 09 19:43:55 crc kubenswrapper[4750]: I0309 19:43:55.103478 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhdx5" event={"ID":"eb5bc9f7-34d2-4180-8633-5deac70271d8","Type":"ContainerStarted","Data":"72ae21e296e3beb7be7b7880a29faef7607e18aff276e8921821ad66947a6e70"} Mar 09 19:43:55 crc kubenswrapper[4750]: I0309 19:43:55.125212 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rsmwj" podStartSLOduration=3.840548773 podStartE2EDuration="13.125192595s" podCreationTimestamp="2026-03-09 19:43:42 +0000 UTC" firstStartedPulling="2026-03-09 19:43:44.972747631 +0000 UTC m=+4706.315220039" lastFinishedPulling="2026-03-09 19:43:54.257391463 +0000 UTC m=+4715.599863861" observedRunningTime="2026-03-09 19:43:55.120742274 +0000 UTC m=+4716.463214672" watchObservedRunningTime="2026-03-09 19:43:55.125192595 +0000 UTC m=+4716.467664993" Mar 09 19:43:57 crc kubenswrapper[4750]: I0309 19:43:57.122208 4750 generic.go:334] "Generic (PLEG): container finished" podID="eb5bc9f7-34d2-4180-8633-5deac70271d8" containerID="72ae21e296e3beb7be7b7880a29faef7607e18aff276e8921821ad66947a6e70" exitCode=0 Mar 09 19:43:57 crc kubenswrapper[4750]: I0309 19:43:57.122246 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhdx5" event={"ID":"eb5bc9f7-34d2-4180-8633-5deac70271d8","Type":"ContainerDied","Data":"72ae21e296e3beb7be7b7880a29faef7607e18aff276e8921821ad66947a6e70"} Mar 09 19:43:58 crc kubenswrapper[4750]: I0309 19:43:58.136549 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhdx5" event={"ID":"eb5bc9f7-34d2-4180-8633-5deac70271d8","Type":"ContainerStarted","Data":"cc30c62d16f84bc0fc4c2c27e050af225547012db6c0f9c4478bce1700ec3425"} Mar 09 19:43:58 crc kubenswrapper[4750]: I0309 19:43:58.166539 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bhdx5" podStartSLOduration=3.379735683 podStartE2EDuration="20.166492062s" podCreationTimestamp="2026-03-09 19:43:38 +0000 UTC" firstStartedPulling="2026-03-09 19:43:40.928997951 +0000 UTC m=+4702.271470389" lastFinishedPulling="2026-03-09 19:43:57.71575436 +0000 UTC m=+4719.058226768" observedRunningTime="2026-03-09 19:43:58.157730785 +0000 UTC m=+4719.500203183" watchObservedRunningTime="2026-03-09 19:43:58.166492062 +0000 UTC m=+4719.508964460" Mar 09 19:43:59 crc kubenswrapper[4750]: I0309 19:43:59.345428 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:43:59 crc kubenswrapper[4750]: I0309 19:43:59.345753 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.181709 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551424-bnw9n"] Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.184140 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551424-bnw9n" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.187301 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.187568 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.188003 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.219687 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551424-bnw9n"] Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.335567 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwhfg\" (UniqueName: \"kubernetes.io/projected/0822dfea-565b-4508-a995-83a244b0c492-kube-api-access-kwhfg\") pod \"auto-csr-approver-29551424-bnw9n\" (UID: \"0822dfea-565b-4508-a995-83a244b0c492\") " pod="openshift-infra/auto-csr-approver-29551424-bnw9n" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.417032 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bhdx5" podUID="eb5bc9f7-34d2-4180-8633-5deac70271d8" containerName="registry-server" probeResult="failure" output=< Mar 09 19:44:00 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:44:00 crc kubenswrapper[4750]: > Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.437733 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwhfg\" (UniqueName: \"kubernetes.io/projected/0822dfea-565b-4508-a995-83a244b0c492-kube-api-access-kwhfg\") pod \"auto-csr-approver-29551424-bnw9n\" (UID: \"0822dfea-565b-4508-a995-83a244b0c492\") " pod="openshift-infra/auto-csr-approver-29551424-bnw9n" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.460437 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwhfg\" (UniqueName: \"kubernetes.io/projected/0822dfea-565b-4508-a995-83a244b0c492-kube-api-access-kwhfg\") pod \"auto-csr-approver-29551424-bnw9n\" (UID: \"0822dfea-565b-4508-a995-83a244b0c492\") " pod="openshift-infra/auto-csr-approver-29551424-bnw9n" Mar 09 19:44:00 crc kubenswrapper[4750]: I0309 19:44:00.509439 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551424-bnw9n" Mar 09 19:44:01 crc kubenswrapper[4750]: W0309 19:44:01.057389 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0822dfea_565b_4508_a995_83a244b0c492.slice/crio-4b9d9772ccbc45e6975157dbd55d1f3fda0eed9b574722fe9abc73b3dbf656ac WatchSource:0}: Error finding container 4b9d9772ccbc45e6975157dbd55d1f3fda0eed9b574722fe9abc73b3dbf656ac: Status 404 returned error can't find the container with id 4b9d9772ccbc45e6975157dbd55d1f3fda0eed9b574722fe9abc73b3dbf656ac Mar 09 19:44:01 crc kubenswrapper[4750]: I0309 19:44:01.057596 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551424-bnw9n"] Mar 09 19:44:01 crc kubenswrapper[4750]: I0309 19:44:01.183993 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551424-bnw9n" event={"ID":"0822dfea-565b-4508-a995-83a244b0c492","Type":"ContainerStarted","Data":"4b9d9772ccbc45e6975157dbd55d1f3fda0eed9b574722fe9abc73b3dbf656ac"} Mar 09 19:44:02 crc kubenswrapper[4750]: I0309 19:44:02.720534 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:44:02 crc kubenswrapper[4750]: I0309 19:44:02.721849 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:44:03 crc kubenswrapper[4750]: I0309 19:44:03.209926 4750 generic.go:334] "Generic (PLEG): container finished" podID="0822dfea-565b-4508-a995-83a244b0c492" containerID="a920df654f36f88feb468391f36876124b6942f5bd33e934e944a3903e68a669" exitCode=0 Mar 09 19:44:03 crc kubenswrapper[4750]: I0309 19:44:03.210032 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551424-bnw9n" event={"ID":"0822dfea-565b-4508-a995-83a244b0c492","Type":"ContainerDied","Data":"a920df654f36f88feb468391f36876124b6942f5bd33e934e944a3903e68a669"} Mar 09 19:44:03 crc kubenswrapper[4750]: I0309 19:44:03.798998 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rsmwj" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="registry-server" probeResult="failure" output=< Mar 09 19:44:03 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:44:03 crc kubenswrapper[4750]: > Mar 09 19:44:04 crc kubenswrapper[4750]: I0309 19:44:04.373747 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:44:04 crc kubenswrapper[4750]: E0309 19:44:04.374719 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:44:04 crc kubenswrapper[4750]: I0309 19:44:04.988052 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551424-bnw9n" Mar 09 19:44:05 crc kubenswrapper[4750]: I0309 19:44:05.164030 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwhfg\" (UniqueName: \"kubernetes.io/projected/0822dfea-565b-4508-a995-83a244b0c492-kube-api-access-kwhfg\") pod \"0822dfea-565b-4508-a995-83a244b0c492\" (UID: \"0822dfea-565b-4508-a995-83a244b0c492\") " Mar 09 19:44:05 crc kubenswrapper[4750]: I0309 19:44:05.174579 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0822dfea-565b-4508-a995-83a244b0c492-kube-api-access-kwhfg" (OuterVolumeSpecName: "kube-api-access-kwhfg") pod "0822dfea-565b-4508-a995-83a244b0c492" (UID: "0822dfea-565b-4508-a995-83a244b0c492"). InnerVolumeSpecName "kube-api-access-kwhfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:44:05 crc kubenswrapper[4750]: I0309 19:44:05.234180 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551424-bnw9n" event={"ID":"0822dfea-565b-4508-a995-83a244b0c492","Type":"ContainerDied","Data":"4b9d9772ccbc45e6975157dbd55d1f3fda0eed9b574722fe9abc73b3dbf656ac"} Mar 09 19:44:05 crc kubenswrapper[4750]: I0309 19:44:05.234238 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b9d9772ccbc45e6975157dbd55d1f3fda0eed9b574722fe9abc73b3dbf656ac" Mar 09 19:44:05 crc kubenswrapper[4750]: I0309 19:44:05.234299 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551424-bnw9n" Mar 09 19:44:05 crc kubenswrapper[4750]: I0309 19:44:05.267379 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwhfg\" (UniqueName: \"kubernetes.io/projected/0822dfea-565b-4508-a995-83a244b0c492-kube-api-access-kwhfg\") on node \"crc\" DevicePath \"\"" Mar 09 19:44:05 crc kubenswrapper[4750]: I0309 19:44:05.373543 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:44:05 crc kubenswrapper[4750]: E0309 19:44:05.374109 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:44:05 crc kubenswrapper[4750]: E0309 19:44:05.435447 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0822dfea_565b_4508_a995_83a244b0c492.slice\": RecentStats: unable to find data in memory cache]" Mar 09 19:44:06 crc kubenswrapper[4750]: I0309 19:44:06.066407 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551418-9l4wb"] Mar 09 19:44:06 crc kubenswrapper[4750]: I0309 19:44:06.075178 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551418-9l4wb"] Mar 09 19:44:07 crc kubenswrapper[4750]: I0309 19:44:07.383150 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="675d07ed-482d-4b1e-b5ab-2553e5a64214" path="/var/lib/kubelet/pods/675d07ed-482d-4b1e-b5ab-2553e5a64214/volumes" Mar 09 19:44:09 crc kubenswrapper[4750]: I0309 19:44:09.409626 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:44:09 crc kubenswrapper[4750]: I0309 19:44:09.470893 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bhdx5" Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.035366 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bhdx5"] Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.211555 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qsrwr"] Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.211998 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qsrwr" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="registry-server" containerID="cri-o://a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784" gracePeriod=2 Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.742783 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.789474 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mtrg\" (UniqueName: \"kubernetes.io/projected/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-kube-api-access-6mtrg\") pod \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.789619 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-catalog-content\") pod \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.789710 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-utilities\") pod \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\" (UID: \"cdd878ef-1b79-49cd-a6bb-b756449fc7d8\") " Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.790391 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-utilities" (OuterVolumeSpecName: "utilities") pod "cdd878ef-1b79-49cd-a6bb-b756449fc7d8" (UID: "cdd878ef-1b79-49cd-a6bb-b756449fc7d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.795424 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-kube-api-access-6mtrg" (OuterVolumeSpecName: "kube-api-access-6mtrg") pod "cdd878ef-1b79-49cd-a6bb-b756449fc7d8" (UID: "cdd878ef-1b79-49cd-a6bb-b756449fc7d8"). InnerVolumeSpecName "kube-api-access-6mtrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.892338 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mtrg\" (UniqueName: \"kubernetes.io/projected/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-kube-api-access-6mtrg\") on node \"crc\" DevicePath \"\"" Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.892660 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.910262 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdd878ef-1b79-49cd-a6bb-b756449fc7d8" (UID: "cdd878ef-1b79-49cd-a6bb-b756449fc7d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:44:10 crc kubenswrapper[4750]: I0309 19:44:10.995104 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd878ef-1b79-49cd-a6bb-b756449fc7d8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.298897 4750 generic.go:334] "Generic (PLEG): container finished" podID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerID="a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784" exitCode=0 Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.299870 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsrwr" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.301779 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsrwr" event={"ID":"cdd878ef-1b79-49cd-a6bb-b756449fc7d8","Type":"ContainerDied","Data":"a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784"} Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.302097 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsrwr" event={"ID":"cdd878ef-1b79-49cd-a6bb-b756449fc7d8","Type":"ContainerDied","Data":"002bb66edcf6f6d7dddf4733fd8e55e8611d2cf193cb2ebb587b044fe496d820"} Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.302155 4750 scope.go:117] "RemoveContainer" containerID="a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.324608 4750 scope.go:117] "RemoveContainer" containerID="b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.337619 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qsrwr"] Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.348793 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qsrwr"] Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.365416 4750 scope.go:117] "RemoveContainer" containerID="f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.384103 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" path="/var/lib/kubelet/pods/cdd878ef-1b79-49cd-a6bb-b756449fc7d8/volumes" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.389676 4750 scope.go:117] "RemoveContainer" containerID="a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784" Mar 09 19:44:11 crc kubenswrapper[4750]: E0309 19:44:11.390160 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784\": container with ID starting with a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784 not found: ID does not exist" containerID="a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.390190 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784"} err="failed to get container status \"a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784\": rpc error: code = NotFound desc = could not find container \"a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784\": container with ID starting with a18d094af6ed0b65ef116c81b2ce815cd6bcbb04dac8f0c0a9178a43e9470784 not found: ID does not exist" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.390212 4750 scope.go:117] "RemoveContainer" containerID="b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858" Mar 09 19:44:11 crc kubenswrapper[4750]: E0309 19:44:11.390550 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858\": container with ID starting with b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858 not found: ID does not exist" containerID="b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.390571 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858"} err="failed to get container status \"b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858\": rpc error: code = NotFound desc = could not find container \"b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858\": container with ID starting with b4f10107e61ea20ae175151dba9c49a0dc9404303f6084a9ef9e71ac5c977858 not found: ID does not exist" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.390584 4750 scope.go:117] "RemoveContainer" containerID="f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491" Mar 09 19:44:11 crc kubenswrapper[4750]: E0309 19:44:11.392210 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491\": container with ID starting with f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491 not found: ID does not exist" containerID="f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491" Mar 09 19:44:11 crc kubenswrapper[4750]: I0309 19:44:11.392339 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491"} err="failed to get container status \"f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491\": rpc error: code = NotFound desc = could not find container \"f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491\": container with ID starting with f4f3a1bad1d41b2f1704cae4f4d1f742b87054b83a38949a5e3e0b23576e0491 not found: ID does not exist" Mar 09 19:44:12 crc kubenswrapper[4750]: I0309 19:44:12.789865 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:44:12 crc kubenswrapper[4750]: I0309 19:44:12.846388 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:44:14 crc kubenswrapper[4750]: I0309 19:44:14.605553 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsmwj"] Mar 09 19:44:14 crc kubenswrapper[4750]: I0309 19:44:14.606123 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rsmwj" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="registry-server" containerID="cri-o://e79cc5318b76c037202f6c41fc0f1eba87af9af5ac3743eccc42e8d9e555451a" gracePeriod=2 Mar 09 19:44:15 crc kubenswrapper[4750]: I0309 19:44:15.343685 4750 generic.go:334] "Generic (PLEG): container finished" podID="da237bf5-477b-4725-9388-55c35bcbfd30" containerID="e79cc5318b76c037202f6c41fc0f1eba87af9af5ac3743eccc42e8d9e555451a" exitCode=0 Mar 09 19:44:15 crc kubenswrapper[4750]: I0309 19:44:15.343740 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsmwj" event={"ID":"da237bf5-477b-4725-9388-55c35bcbfd30","Type":"ContainerDied","Data":"e79cc5318b76c037202f6c41fc0f1eba87af9af5ac3743eccc42e8d9e555451a"} Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.182019 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.243764 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh9p5\" (UniqueName: \"kubernetes.io/projected/da237bf5-477b-4725-9388-55c35bcbfd30-kube-api-access-hh9p5\") pod \"da237bf5-477b-4725-9388-55c35bcbfd30\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.243821 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-catalog-content\") pod \"da237bf5-477b-4725-9388-55c35bcbfd30\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.243850 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-utilities\") pod \"da237bf5-477b-4725-9388-55c35bcbfd30\" (UID: \"da237bf5-477b-4725-9388-55c35bcbfd30\") " Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.244981 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-utilities" (OuterVolumeSpecName: "utilities") pod "da237bf5-477b-4725-9388-55c35bcbfd30" (UID: "da237bf5-477b-4725-9388-55c35bcbfd30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.250597 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da237bf5-477b-4725-9388-55c35bcbfd30-kube-api-access-hh9p5" (OuterVolumeSpecName: "kube-api-access-hh9p5") pod "da237bf5-477b-4725-9388-55c35bcbfd30" (UID: "da237bf5-477b-4725-9388-55c35bcbfd30"). InnerVolumeSpecName "kube-api-access-hh9p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.311528 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da237bf5-477b-4725-9388-55c35bcbfd30" (UID: "da237bf5-477b-4725-9388-55c35bcbfd30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.346019 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh9p5\" (UniqueName: \"kubernetes.io/projected/da237bf5-477b-4725-9388-55c35bcbfd30-kube-api-access-hh9p5\") on node \"crc\" DevicePath \"\"" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.346102 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.346117 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da237bf5-477b-4725-9388-55c35bcbfd30-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.359510 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsmwj" event={"ID":"da237bf5-477b-4725-9388-55c35bcbfd30","Type":"ContainerDied","Data":"8f2fa6cca48976af1e4d2bcb175f9c114cc0c0e0bf06f685b254a8baebc37536"} Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.359571 4750 scope.go:117] "RemoveContainer" containerID="e79cc5318b76c037202f6c41fc0f1eba87af9af5ac3743eccc42e8d9e555451a" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.359592 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsmwj" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.374662 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:44:16 crc kubenswrapper[4750]: E0309 19:44:16.375391 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.384012 4750 scope.go:117] "RemoveContainer" containerID="94c137630b437b4249dcc3bba6e61968b9a530aa015fa4e1c8a55196fa35961a" Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.399761 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsmwj"] Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.410741 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rsmwj"] Mar 09 19:44:16 crc kubenswrapper[4750]: I0309 19:44:16.433673 4750 scope.go:117] "RemoveContainer" containerID="a13d174f807bfd6825f31d04628faeacb781f227114544fa900803fe9beb17ee" Mar 09 19:44:17 crc kubenswrapper[4750]: I0309 19:44:17.390430 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" path="/var/lib/kubelet/pods/da237bf5-477b-4725-9388-55c35bcbfd30/volumes" Mar 09 19:44:19 crc kubenswrapper[4750]: I0309 19:44:19.390278 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:44:19 crc kubenswrapper[4750]: E0309 19:44:19.390978 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:44:29 crc kubenswrapper[4750]: I0309 19:44:29.380907 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:44:29 crc kubenswrapper[4750]: E0309 19:44:29.381663 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:44:30 crc kubenswrapper[4750]: I0309 19:44:30.374214 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:44:30 crc kubenswrapper[4750]: E0309 19:44:30.374615 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:44:43 crc kubenswrapper[4750]: I0309 19:44:43.374118 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:44:43 crc kubenswrapper[4750]: E0309 19:44:43.374947 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:44:44 crc kubenswrapper[4750]: I0309 19:44:44.374378 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:44:44 crc kubenswrapper[4750]: E0309 19:44:44.374660 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:44:49 crc kubenswrapper[4750]: I0309 19:44:49.177200 4750 scope.go:117] "RemoveContainer" containerID="b1aa1aa93d21843cd6efea4bc6c5faab0b92a8c0f5b2b4452c0350f66c918542" Mar 09 19:44:55 crc kubenswrapper[4750]: I0309 19:44:55.374760 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:44:55 crc kubenswrapper[4750]: E0309 19:44:55.375543 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:44:56 crc kubenswrapper[4750]: I0309 19:44:56.374281 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:44:56 crc kubenswrapper[4750]: E0309 19:44:56.374568 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.184705 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s"] Mar 09 19:45:00 crc kubenswrapper[4750]: E0309 19:45:00.185887 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0822dfea-565b-4508-a995-83a244b0c492" containerName="oc" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.185908 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="0822dfea-565b-4508-a995-83a244b0c492" containerName="oc" Mar 09 19:45:00 crc kubenswrapper[4750]: E0309 19:45:00.185934 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="extract-utilities" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.185944 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="extract-utilities" Mar 09 19:45:00 crc kubenswrapper[4750]: E0309 19:45:00.185974 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="registry-server" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.185986 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="registry-server" Mar 09 19:45:00 crc kubenswrapper[4750]: E0309 19:45:00.186000 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="extract-utilities" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.186009 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="extract-utilities" Mar 09 19:45:00 crc kubenswrapper[4750]: E0309 19:45:00.186032 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="registry-server" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.186041 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="registry-server" Mar 09 19:45:00 crc kubenswrapper[4750]: E0309 19:45:00.186061 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="extract-content" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.186070 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="extract-content" Mar 09 19:45:00 crc kubenswrapper[4750]: E0309 19:45:00.186107 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="extract-content" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.186117 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="extract-content" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.186445 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="0822dfea-565b-4508-a995-83a244b0c492" containerName="oc" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.186476 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd878ef-1b79-49cd-a6bb-b756449fc7d8" containerName="registry-server" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.186497 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="da237bf5-477b-4725-9388-55c35bcbfd30" containerName="registry-server" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.187562 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.192159 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-secret-volume\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.192247 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.192485 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cs79\" (UniqueName: \"kubernetes.io/projected/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-kube-api-access-6cs79\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.192596 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-config-volume\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.193023 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.206717 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s"] Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.293944 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cs79\" (UniqueName: \"kubernetes.io/projected/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-kube-api-access-6cs79\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.294008 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-config-volume\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.294056 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-secret-volume\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.295566 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-config-volume\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.301899 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-secret-volume\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.322124 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cs79\" (UniqueName: \"kubernetes.io/projected/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-kube-api-access-6cs79\") pod \"collect-profiles-29551425-4q28s\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:00 crc kubenswrapper[4750]: I0309 19:45:00.531467 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:01 crc kubenswrapper[4750]: I0309 19:45:01.053434 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s"] Mar 09 19:45:01 crc kubenswrapper[4750]: I0309 19:45:01.871924 4750 generic.go:334] "Generic (PLEG): container finished" podID="1e5f17c9-8b51-48f5-9d97-2ece1bc135a3" containerID="ec7d73d825bd641c0c0ad0418a22e3b9aeb22d633027a07300932d3befe583b8" exitCode=0 Mar 09 19:45:01 crc kubenswrapper[4750]: I0309 19:45:01.872017 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" event={"ID":"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3","Type":"ContainerDied","Data":"ec7d73d825bd641c0c0ad0418a22e3b9aeb22d633027a07300932d3befe583b8"} Mar 09 19:45:01 crc kubenswrapper[4750]: I0309 19:45:01.872230 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" event={"ID":"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3","Type":"ContainerStarted","Data":"6559ff0537a1de95c2c23a90e4f67fdcd3270f01c2c41e3b55f2cd74efa7fe63"} Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.291231 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.485303 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-secret-volume\") pod \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.485458 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cs79\" (UniqueName: \"kubernetes.io/projected/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-kube-api-access-6cs79\") pod \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.485591 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-config-volume\") pod \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\" (UID: \"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3\") " Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.488207 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-config-volume" (OuterVolumeSpecName: "config-volume") pod "1e5f17c9-8b51-48f5-9d97-2ece1bc135a3" (UID: "1e5f17c9-8b51-48f5-9d97-2ece1bc135a3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.492415 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-kube-api-access-6cs79" (OuterVolumeSpecName: "kube-api-access-6cs79") pod "1e5f17c9-8b51-48f5-9d97-2ece1bc135a3" (UID: "1e5f17c9-8b51-48f5-9d97-2ece1bc135a3"). InnerVolumeSpecName "kube-api-access-6cs79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.493330 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1e5f17c9-8b51-48f5-9d97-2ece1bc135a3" (UID: "1e5f17c9-8b51-48f5-9d97-2ece1bc135a3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.589187 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cs79\" (UniqueName: \"kubernetes.io/projected/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-kube-api-access-6cs79\") on node \"crc\" DevicePath \"\"" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.589675 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.589766 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e5f17c9-8b51-48f5-9d97-2ece1bc135a3-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.907543 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" event={"ID":"1e5f17c9-8b51-48f5-9d97-2ece1bc135a3","Type":"ContainerDied","Data":"6559ff0537a1de95c2c23a90e4f67fdcd3270f01c2c41e3b55f2cd74efa7fe63"} Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.907596 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551425-4q28s" Mar 09 19:45:03 crc kubenswrapper[4750]: I0309 19:45:03.907617 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6559ff0537a1de95c2c23a90e4f67fdcd3270f01c2c41e3b55f2cd74efa7fe63" Mar 09 19:45:04 crc kubenswrapper[4750]: I0309 19:45:04.382507 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm"] Mar 09 19:45:04 crc kubenswrapper[4750]: I0309 19:45:04.394882 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551380-h4ckm"] Mar 09 19:45:05 crc kubenswrapper[4750]: I0309 19:45:05.399144 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b726a609-8d2b-406e-8f7a-1e37d2c8a7a5" path="/var/lib/kubelet/pods/b726a609-8d2b-406e-8f7a-1e37d2c8a7a5/volumes" Mar 09 19:45:08 crc kubenswrapper[4750]: I0309 19:45:08.374487 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:45:08 crc kubenswrapper[4750]: I0309 19:45:08.375405 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:45:08 crc kubenswrapper[4750]: E0309 19:45:08.375722 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:45:08 crc kubenswrapper[4750]: E0309 19:45:08.376102 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:45:19 crc kubenswrapper[4750]: I0309 19:45:19.389244 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:45:19 crc kubenswrapper[4750]: E0309 19:45:19.390296 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:45:20 crc kubenswrapper[4750]: I0309 19:45:20.374710 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:45:20 crc kubenswrapper[4750]: E0309 19:45:20.375266 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:45:33 crc kubenswrapper[4750]: I0309 19:45:33.374683 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:45:33 crc kubenswrapper[4750]: I0309 19:45:33.375700 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:45:33 crc kubenswrapper[4750]: E0309 19:45:33.376353 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:45:34 crc kubenswrapper[4750]: I0309 19:45:34.310146 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188"} Mar 09 19:45:35 crc kubenswrapper[4750]: I0309 19:45:35.189411 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:45:36 crc kubenswrapper[4750]: I0309 19:45:36.340549 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" exitCode=1 Mar 09 19:45:36 crc kubenswrapper[4750]: I0309 19:45:36.340680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188"} Mar 09 19:45:36 crc kubenswrapper[4750]: I0309 19:45:36.341213 4750 scope.go:117] "RemoveContainer" containerID="fbc7473130b072e832a36cff0e2808cc359f8ae8d6112e0a34ab683c3f57dfa3" Mar 09 19:45:36 crc kubenswrapper[4750]: I0309 19:45:36.342386 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:45:36 crc kubenswrapper[4750]: E0309 19:45:36.343464 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:45:38 crc kubenswrapper[4750]: I0309 19:45:38.189715 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:45:38 crc kubenswrapper[4750]: I0309 19:45:38.191014 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:45:38 crc kubenswrapper[4750]: E0309 19:45:38.191667 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:45:40 crc kubenswrapper[4750]: I0309 19:45:40.189523 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:45:40 crc kubenswrapper[4750]: I0309 19:45:40.192607 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:45:40 crc kubenswrapper[4750]: E0309 19:45:40.193421 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:45:43 crc kubenswrapper[4750]: I0309 19:45:43.975801 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vzrx6"] Mar 09 19:45:43 crc kubenswrapper[4750]: E0309 19:45:43.976782 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e5f17c9-8b51-48f5-9d97-2ece1bc135a3" containerName="collect-profiles" Mar 09 19:45:43 crc kubenswrapper[4750]: I0309 19:45:43.976804 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e5f17c9-8b51-48f5-9d97-2ece1bc135a3" containerName="collect-profiles" Mar 09 19:45:43 crc kubenswrapper[4750]: I0309 19:45:43.977266 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e5f17c9-8b51-48f5-9d97-2ece1bc135a3" containerName="collect-profiles" Mar 09 19:45:43 crc kubenswrapper[4750]: I0309 19:45:43.980074 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:43 crc kubenswrapper[4750]: I0309 19:45:43.987039 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzrx6"] Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.115750 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-utilities\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.115815 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4nsh\" (UniqueName: \"kubernetes.io/projected/6d8a1199-a8d0-4a0e-8849-1314a2267a76-kube-api-access-x4nsh\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.115967 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-catalog-content\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.218251 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-catalog-content\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.218411 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-utilities\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.218450 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4nsh\" (UniqueName: \"kubernetes.io/projected/6d8a1199-a8d0-4a0e-8849-1314a2267a76-kube-api-access-x4nsh\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.218884 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-catalog-content\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.218898 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-utilities\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.249114 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4nsh\" (UniqueName: \"kubernetes.io/projected/6d8a1199-a8d0-4a0e-8849-1314a2267a76-kube-api-access-x4nsh\") pod \"redhat-marketplace-vzrx6\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.304116 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:44 crc kubenswrapper[4750]: I0309 19:45:44.824279 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzrx6"] Mar 09 19:45:45 crc kubenswrapper[4750]: I0309 19:45:45.449089 4750 generic.go:334] "Generic (PLEG): container finished" podID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerID="f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe" exitCode=0 Mar 09 19:45:45 crc kubenswrapper[4750]: I0309 19:45:45.449193 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzrx6" event={"ID":"6d8a1199-a8d0-4a0e-8849-1314a2267a76","Type":"ContainerDied","Data":"f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe"} Mar 09 19:45:45 crc kubenswrapper[4750]: I0309 19:45:45.449420 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzrx6" event={"ID":"6d8a1199-a8d0-4a0e-8849-1314a2267a76","Type":"ContainerStarted","Data":"20cc655c2008f5767382b7009f07ee8ee1dd189745a75d6cd499a5b2c675e4dd"} Mar 09 19:45:46 crc kubenswrapper[4750]: I0309 19:45:46.478851 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzrx6" event={"ID":"6d8a1199-a8d0-4a0e-8849-1314a2267a76","Type":"ContainerStarted","Data":"a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e"} Mar 09 19:45:47 crc kubenswrapper[4750]: I0309 19:45:47.373884 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:45:47 crc kubenswrapper[4750]: E0309 19:45:47.374592 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:45:47 crc kubenswrapper[4750]: I0309 19:45:47.496699 4750 generic.go:334] "Generic (PLEG): container finished" podID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerID="a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e" exitCode=0 Mar 09 19:45:47 crc kubenswrapper[4750]: I0309 19:45:47.496743 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzrx6" event={"ID":"6d8a1199-a8d0-4a0e-8849-1314a2267a76","Type":"ContainerDied","Data":"a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e"} Mar 09 19:45:48 crc kubenswrapper[4750]: I0309 19:45:48.509193 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzrx6" event={"ID":"6d8a1199-a8d0-4a0e-8849-1314a2267a76","Type":"ContainerStarted","Data":"f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7"} Mar 09 19:45:48 crc kubenswrapper[4750]: I0309 19:45:48.533098 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vzrx6" podStartSLOduration=3.017809173 podStartE2EDuration="5.533076466s" podCreationTimestamp="2026-03-09 19:45:43 +0000 UTC" firstStartedPulling="2026-03-09 19:45:45.45049058 +0000 UTC m=+4826.792962978" lastFinishedPulling="2026-03-09 19:45:47.965757843 +0000 UTC m=+4829.308230271" observedRunningTime="2026-03-09 19:45:48.526245232 +0000 UTC m=+4829.868717650" watchObservedRunningTime="2026-03-09 19:45:48.533076466 +0000 UTC m=+4829.875548874" Mar 09 19:45:49 crc kubenswrapper[4750]: I0309 19:45:49.372950 4750 scope.go:117] "RemoveContainer" containerID="3e0918883adb7aa11f8553ebdd0cf2413b90ef754e1bffcbc4c47b2b9c38dc75" Mar 09 19:45:52 crc kubenswrapper[4750]: I0309 19:45:52.374031 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:45:52 crc kubenswrapper[4750]: E0309 19:45:52.375194 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:45:54 crc kubenswrapper[4750]: I0309 19:45:54.305164 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:54 crc kubenswrapper[4750]: I0309 19:45:54.305244 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:54 crc kubenswrapper[4750]: I0309 19:45:54.461872 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:54 crc kubenswrapper[4750]: I0309 19:45:54.676716 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:54 crc kubenswrapper[4750]: I0309 19:45:54.739606 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzrx6"] Mar 09 19:45:56 crc kubenswrapper[4750]: I0309 19:45:56.621794 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vzrx6" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="registry-server" containerID="cri-o://f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7" gracePeriod=2 Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.263947 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.353856 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-catalog-content\") pod \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.354062 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-utilities\") pod \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.354151 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4nsh\" (UniqueName: \"kubernetes.io/projected/6d8a1199-a8d0-4a0e-8849-1314a2267a76-kube-api-access-x4nsh\") pod \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\" (UID: \"6d8a1199-a8d0-4a0e-8849-1314a2267a76\") " Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.355493 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-utilities" (OuterVolumeSpecName: "utilities") pod "6d8a1199-a8d0-4a0e-8849-1314a2267a76" (UID: "6d8a1199-a8d0-4a0e-8849-1314a2267a76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.360708 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d8a1199-a8d0-4a0e-8849-1314a2267a76-kube-api-access-x4nsh" (OuterVolumeSpecName: "kube-api-access-x4nsh") pod "6d8a1199-a8d0-4a0e-8849-1314a2267a76" (UID: "6d8a1199-a8d0-4a0e-8849-1314a2267a76"). InnerVolumeSpecName "kube-api-access-x4nsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.394737 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d8a1199-a8d0-4a0e-8849-1314a2267a76" (UID: "6d8a1199-a8d0-4a0e-8849-1314a2267a76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.458166 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.458201 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d8a1199-a8d0-4a0e-8849-1314a2267a76-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.458215 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4nsh\" (UniqueName: \"kubernetes.io/projected/6d8a1199-a8d0-4a0e-8849-1314a2267a76-kube-api-access-x4nsh\") on node \"crc\" DevicePath \"\"" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.638951 4750 generic.go:334] "Generic (PLEG): container finished" podID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerID="f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7" exitCode=0 Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.639058 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzrx6" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.639736 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzrx6" event={"ID":"6d8a1199-a8d0-4a0e-8849-1314a2267a76","Type":"ContainerDied","Data":"f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7"} Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.639808 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzrx6" event={"ID":"6d8a1199-a8d0-4a0e-8849-1314a2267a76","Type":"ContainerDied","Data":"20cc655c2008f5767382b7009f07ee8ee1dd189745a75d6cd499a5b2c675e4dd"} Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.639874 4750 scope.go:117] "RemoveContainer" containerID="f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.670269 4750 scope.go:117] "RemoveContainer" containerID="a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.696466 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzrx6"] Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.715991 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzrx6"] Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.724567 4750 scope.go:117] "RemoveContainer" containerID="f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.769825 4750 scope.go:117] "RemoveContainer" containerID="f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7" Mar 09 19:45:57 crc kubenswrapper[4750]: E0309 19:45:57.770224 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7\": container with ID starting with f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7 not found: ID does not exist" containerID="f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.770290 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7"} err="failed to get container status \"f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7\": rpc error: code = NotFound desc = could not find container \"f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7\": container with ID starting with f5d37291578ed4df1cc0ce3dd555023b92e5b5b540c565d9b20061fb3b33ece7 not found: ID does not exist" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.770330 4750 scope.go:117] "RemoveContainer" containerID="a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e" Mar 09 19:45:57 crc kubenswrapper[4750]: E0309 19:45:57.770693 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e\": container with ID starting with a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e not found: ID does not exist" containerID="a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.770729 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e"} err="failed to get container status \"a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e\": rpc error: code = NotFound desc = could not find container \"a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e\": container with ID starting with a4eda419738bd1a11e684a21a1c00d38c7ae466f29105dc84670e4c26562752e not found: ID does not exist" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.770752 4750 scope.go:117] "RemoveContainer" containerID="f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe" Mar 09 19:45:57 crc kubenswrapper[4750]: E0309 19:45:57.771022 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe\": container with ID starting with f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe not found: ID does not exist" containerID="f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe" Mar 09 19:45:57 crc kubenswrapper[4750]: I0309 19:45:57.771066 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe"} err="failed to get container status \"f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe\": rpc error: code = NotFound desc = could not find container \"f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe\": container with ID starting with f5a3538d7c3f0ac1f0e6fabada4e5e39dde6ae944d89438fd97424d2810310fe not found: ID does not exist" Mar 09 19:45:59 crc kubenswrapper[4750]: I0309 19:45:59.398807 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" path="/var/lib/kubelet/pods/6d8a1199-a8d0-4a0e-8849-1314a2267a76/volumes" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.171344 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551426-v2pm4"] Mar 09 19:46:00 crc kubenswrapper[4750]: E0309 19:46:00.172441 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="extract-content" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.172468 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="extract-content" Mar 09 19:46:00 crc kubenswrapper[4750]: E0309 19:46:00.172505 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="extract-utilities" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.172519 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="extract-utilities" Mar 09 19:46:00 crc kubenswrapper[4750]: E0309 19:46:00.172557 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="registry-server" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.172566 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="registry-server" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.172902 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d8a1199-a8d0-4a0e-8849-1314a2267a76" containerName="registry-server" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.174157 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551426-v2pm4" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.179960 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.180175 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.184189 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.224438 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551426-v2pm4"] Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.327149 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27c56\" (UniqueName: \"kubernetes.io/projected/b4948b4b-7c80-4298-913b-2d741c6b7a84-kube-api-access-27c56\") pod \"auto-csr-approver-29551426-v2pm4\" (UID: \"b4948b4b-7c80-4298-913b-2d741c6b7a84\") " pod="openshift-infra/auto-csr-approver-29551426-v2pm4" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.373751 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:46:00 crc kubenswrapper[4750]: E0309 19:46:00.373982 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.429076 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27c56\" (UniqueName: \"kubernetes.io/projected/b4948b4b-7c80-4298-913b-2d741c6b7a84-kube-api-access-27c56\") pod \"auto-csr-approver-29551426-v2pm4\" (UID: \"b4948b4b-7c80-4298-913b-2d741c6b7a84\") " pod="openshift-infra/auto-csr-approver-29551426-v2pm4" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.451721 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27c56\" (UniqueName: \"kubernetes.io/projected/b4948b4b-7c80-4298-913b-2d741c6b7a84-kube-api-access-27c56\") pod \"auto-csr-approver-29551426-v2pm4\" (UID: \"b4948b4b-7c80-4298-913b-2d741c6b7a84\") " pod="openshift-infra/auto-csr-approver-29551426-v2pm4" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.523166 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551426-v2pm4" Mar 09 19:46:00 crc kubenswrapper[4750]: I0309 19:46:00.816598 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551426-v2pm4"] Mar 09 19:46:00 crc kubenswrapper[4750]: W0309 19:46:00.822891 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4948b4b_7c80_4298_913b_2d741c6b7a84.slice/crio-8cc6a7609436e6204e98fba3c31e5f982e965244ada61afde18ba6680e1ac3ce WatchSource:0}: Error finding container 8cc6a7609436e6204e98fba3c31e5f982e965244ada61afde18ba6680e1ac3ce: Status 404 returned error can't find the container with id 8cc6a7609436e6204e98fba3c31e5f982e965244ada61afde18ba6680e1ac3ce Mar 09 19:46:01 crc kubenswrapper[4750]: I0309 19:46:01.708482 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551426-v2pm4" event={"ID":"b4948b4b-7c80-4298-913b-2d741c6b7a84","Type":"ContainerStarted","Data":"8cc6a7609436e6204e98fba3c31e5f982e965244ada61afde18ba6680e1ac3ce"} Mar 09 19:46:02 crc kubenswrapper[4750]: I0309 19:46:02.725202 4750 generic.go:334] "Generic (PLEG): container finished" podID="b4948b4b-7c80-4298-913b-2d741c6b7a84" containerID="87f8f9255d585d2fc0a288a3e14089e5379e67edf11374f9f039c239a58288f7" exitCode=0 Mar 09 19:46:02 crc kubenswrapper[4750]: I0309 19:46:02.725726 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551426-v2pm4" event={"ID":"b4948b4b-7c80-4298-913b-2d741c6b7a84","Type":"ContainerDied","Data":"87f8f9255d585d2fc0a288a3e14089e5379e67edf11374f9f039c239a58288f7"} Mar 09 19:46:04 crc kubenswrapper[4750]: I0309 19:46:04.748453 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551426-v2pm4" event={"ID":"b4948b4b-7c80-4298-913b-2d741c6b7a84","Type":"ContainerDied","Data":"8cc6a7609436e6204e98fba3c31e5f982e965244ada61afde18ba6680e1ac3ce"} Mar 09 19:46:04 crc kubenswrapper[4750]: I0309 19:46:04.748841 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cc6a7609436e6204e98fba3c31e5f982e965244ada61afde18ba6680e1ac3ce" Mar 09 19:46:04 crc kubenswrapper[4750]: I0309 19:46:04.805903 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551426-v2pm4" Mar 09 19:46:04 crc kubenswrapper[4750]: I0309 19:46:04.933434 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27c56\" (UniqueName: \"kubernetes.io/projected/b4948b4b-7c80-4298-913b-2d741c6b7a84-kube-api-access-27c56\") pod \"b4948b4b-7c80-4298-913b-2d741c6b7a84\" (UID: \"b4948b4b-7c80-4298-913b-2d741c6b7a84\") " Mar 09 19:46:04 crc kubenswrapper[4750]: I0309 19:46:04.952960 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4948b4b-7c80-4298-913b-2d741c6b7a84-kube-api-access-27c56" (OuterVolumeSpecName: "kube-api-access-27c56") pod "b4948b4b-7c80-4298-913b-2d741c6b7a84" (UID: "b4948b4b-7c80-4298-913b-2d741c6b7a84"). InnerVolumeSpecName "kube-api-access-27c56". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:46:05 crc kubenswrapper[4750]: I0309 19:46:05.036829 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27c56\" (UniqueName: \"kubernetes.io/projected/b4948b4b-7c80-4298-913b-2d741c6b7a84-kube-api-access-27c56\") on node \"crc\" DevicePath \"\"" Mar 09 19:46:05 crc kubenswrapper[4750]: I0309 19:46:05.757966 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551426-v2pm4" Mar 09 19:46:05 crc kubenswrapper[4750]: I0309 19:46:05.885474 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551420-mj8jl"] Mar 09 19:46:05 crc kubenswrapper[4750]: I0309 19:46:05.902060 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551420-mj8jl"] Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.373761 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:46:07 crc kubenswrapper[4750]: E0309 19:46:07.374939 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.404050 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4207172d-d888-4a93-bba8-9316e3425d5b" path="/var/lib/kubelet/pods/4207172d-d888-4a93-bba8-9316e3425d5b/volumes" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.405337 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-97w7l"] Mar 09 19:46:07 crc kubenswrapper[4750]: E0309 19:46:07.405869 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4948b4b-7c80-4298-913b-2d741c6b7a84" containerName="oc" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.405895 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4948b4b-7c80-4298-913b-2d741c6b7a84" containerName="oc" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.406674 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4948b4b-7c80-4298-913b-2d741c6b7a84" containerName="oc" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.409952 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.423276 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-97w7l"] Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.495971 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-catalog-content\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.496087 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-utilities\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.496165 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr86t\" (UniqueName: \"kubernetes.io/projected/317e3659-3659-42d6-9b0d-1adc70754e5a-kube-api-access-lr86t\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.599726 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-catalog-content\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.599791 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-utilities\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.599851 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr86t\" (UniqueName: \"kubernetes.io/projected/317e3659-3659-42d6-9b0d-1adc70754e5a-kube-api-access-lr86t\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.600405 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-catalog-content\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.600777 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-utilities\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.623469 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr86t\" (UniqueName: \"kubernetes.io/projected/317e3659-3659-42d6-9b0d-1adc70754e5a-kube-api-access-lr86t\") pod \"certified-operators-97w7l\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:07 crc kubenswrapper[4750]: I0309 19:46:07.748596 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:08 crc kubenswrapper[4750]: I0309 19:46:08.296076 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-97w7l"] Mar 09 19:46:08 crc kubenswrapper[4750]: I0309 19:46:08.796056 4750 generic.go:334] "Generic (PLEG): container finished" podID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerID="8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3" exitCode=0 Mar 09 19:46:08 crc kubenswrapper[4750]: I0309 19:46:08.796102 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97w7l" event={"ID":"317e3659-3659-42d6-9b0d-1adc70754e5a","Type":"ContainerDied","Data":"8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3"} Mar 09 19:46:08 crc kubenswrapper[4750]: I0309 19:46:08.796127 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97w7l" event={"ID":"317e3659-3659-42d6-9b0d-1adc70754e5a","Type":"ContainerStarted","Data":"f9cb96da81b686ac2caaf8630cd8f58b5ad6029816e6cb4a5b8328232607e2dc"} Mar 09 19:46:10 crc kubenswrapper[4750]: I0309 19:46:10.827838 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97w7l" event={"ID":"317e3659-3659-42d6-9b0d-1adc70754e5a","Type":"ContainerStarted","Data":"01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f"} Mar 09 19:46:11 crc kubenswrapper[4750]: I0309 19:46:11.844386 4750 generic.go:334] "Generic (PLEG): container finished" podID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerID="01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f" exitCode=0 Mar 09 19:46:11 crc kubenswrapper[4750]: I0309 19:46:11.844453 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97w7l" event={"ID":"317e3659-3659-42d6-9b0d-1adc70754e5a","Type":"ContainerDied","Data":"01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f"} Mar 09 19:46:13 crc kubenswrapper[4750]: I0309 19:46:13.373704 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:46:13 crc kubenswrapper[4750]: E0309 19:46:13.374399 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:46:13 crc kubenswrapper[4750]: I0309 19:46:13.878718 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97w7l" event={"ID":"317e3659-3659-42d6-9b0d-1adc70754e5a","Type":"ContainerStarted","Data":"fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99"} Mar 09 19:46:13 crc kubenswrapper[4750]: I0309 19:46:13.901843 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-97w7l" podStartSLOduration=2.369025984 podStartE2EDuration="6.901824655s" podCreationTimestamp="2026-03-09 19:46:07 +0000 UTC" firstStartedPulling="2026-03-09 19:46:08.798139573 +0000 UTC m=+4850.140611971" lastFinishedPulling="2026-03-09 19:46:13.330938244 +0000 UTC m=+4854.673410642" observedRunningTime="2026-03-09 19:46:13.898921136 +0000 UTC m=+4855.241393574" watchObservedRunningTime="2026-03-09 19:46:13.901824655 +0000 UTC m=+4855.244297063" Mar 09 19:46:17 crc kubenswrapper[4750]: I0309 19:46:17.748755 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:17 crc kubenswrapper[4750]: I0309 19:46:17.749413 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:17 crc kubenswrapper[4750]: I0309 19:46:17.842288 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:19 crc kubenswrapper[4750]: I0309 19:46:19.391235 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:46:19 crc kubenswrapper[4750]: E0309 19:46:19.392107 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:46:26 crc kubenswrapper[4750]: I0309 19:46:26.373484 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:46:26 crc kubenswrapper[4750]: E0309 19:46:26.374814 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:46:27 crc kubenswrapper[4750]: I0309 19:46:27.813758 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:27 crc kubenswrapper[4750]: I0309 19:46:27.882220 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-97w7l"] Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.055458 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-97w7l" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="registry-server" containerID="cri-o://fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99" gracePeriod=2 Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.538929 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.610100 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-catalog-content\") pod \"317e3659-3659-42d6-9b0d-1adc70754e5a\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.610203 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-utilities\") pod \"317e3659-3659-42d6-9b0d-1adc70754e5a\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.610438 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr86t\" (UniqueName: \"kubernetes.io/projected/317e3659-3659-42d6-9b0d-1adc70754e5a-kube-api-access-lr86t\") pod \"317e3659-3659-42d6-9b0d-1adc70754e5a\" (UID: \"317e3659-3659-42d6-9b0d-1adc70754e5a\") " Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.611574 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-utilities" (OuterVolumeSpecName: "utilities") pod "317e3659-3659-42d6-9b0d-1adc70754e5a" (UID: "317e3659-3659-42d6-9b0d-1adc70754e5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.625334 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/317e3659-3659-42d6-9b0d-1adc70754e5a-kube-api-access-lr86t" (OuterVolumeSpecName: "kube-api-access-lr86t") pod "317e3659-3659-42d6-9b0d-1adc70754e5a" (UID: "317e3659-3659-42d6-9b0d-1adc70754e5a"). InnerVolumeSpecName "kube-api-access-lr86t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.677250 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "317e3659-3659-42d6-9b0d-1adc70754e5a" (UID: "317e3659-3659-42d6-9b0d-1adc70754e5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.713237 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.713312 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317e3659-3659-42d6-9b0d-1adc70754e5a-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:46:28 crc kubenswrapper[4750]: I0309 19:46:28.713330 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr86t\" (UniqueName: \"kubernetes.io/projected/317e3659-3659-42d6-9b0d-1adc70754e5a-kube-api-access-lr86t\") on node \"crc\" DevicePath \"\"" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.070768 4750 generic.go:334] "Generic (PLEG): container finished" podID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerID="fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99" exitCode=0 Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.070852 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97w7l" event={"ID":"317e3659-3659-42d6-9b0d-1adc70754e5a","Type":"ContainerDied","Data":"fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99"} Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.070903 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97w7l" event={"ID":"317e3659-3659-42d6-9b0d-1adc70754e5a","Type":"ContainerDied","Data":"f9cb96da81b686ac2caaf8630cd8f58b5ad6029816e6cb4a5b8328232607e2dc"} Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.070938 4750 scope.go:117] "RemoveContainer" containerID="fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.071184 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97w7l" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.109832 4750 scope.go:117] "RemoveContainer" containerID="01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.137672 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-97w7l"] Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.155453 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-97w7l"] Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.168810 4750 scope.go:117] "RemoveContainer" containerID="8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.221731 4750 scope.go:117] "RemoveContainer" containerID="fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99" Mar 09 19:46:29 crc kubenswrapper[4750]: E0309 19:46:29.222375 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99\": container with ID starting with fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99 not found: ID does not exist" containerID="fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.222496 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99"} err="failed to get container status \"fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99\": rpc error: code = NotFound desc = could not find container \"fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99\": container with ID starting with fe9effa635f19ccb5668e8e8f1973317b732831a12d4792e6dd5b5d030aaaa99 not found: ID does not exist" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.222569 4750 scope.go:117] "RemoveContainer" containerID="01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f" Mar 09 19:46:29 crc kubenswrapper[4750]: E0309 19:46:29.223225 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f\": container with ID starting with 01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f not found: ID does not exist" containerID="01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.223271 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f"} err="failed to get container status \"01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f\": rpc error: code = NotFound desc = could not find container \"01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f\": container with ID starting with 01a97a16a1394f61ff451991fe22fa8d644a788c5007ddeef15a977a1874241f not found: ID does not exist" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.223302 4750 scope.go:117] "RemoveContainer" containerID="8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3" Mar 09 19:46:29 crc kubenswrapper[4750]: E0309 19:46:29.223770 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3\": container with ID starting with 8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3 not found: ID does not exist" containerID="8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.223827 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3"} err="failed to get container status \"8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3\": rpc error: code = NotFound desc = could not find container \"8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3\": container with ID starting with 8a7e96d1066193cd394e956232e6a1fe70e07e07b85816f7bc01ab635070c6b3 not found: ID does not exist" Mar 09 19:46:29 crc kubenswrapper[4750]: I0309 19:46:29.393899 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" path="/var/lib/kubelet/pods/317e3659-3659-42d6-9b0d-1adc70754e5a/volumes" Mar 09 19:46:31 crc kubenswrapper[4750]: I0309 19:46:31.373610 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:46:31 crc kubenswrapper[4750]: E0309 19:46:31.374272 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:46:39 crc kubenswrapper[4750]: I0309 19:46:39.383083 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:46:39 crc kubenswrapper[4750]: E0309 19:46:39.384056 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:46:43 crc kubenswrapper[4750]: I0309 19:46:43.374181 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:46:43 crc kubenswrapper[4750]: E0309 19:46:43.375028 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:46:49 crc kubenswrapper[4750]: I0309 19:46:49.463258 4750 scope.go:117] "RemoveContainer" containerID="539cb87d4b09c1b2e472859625d392ec70d7f767301a563af43dcffa3ba1aae7" Mar 09 19:46:51 crc kubenswrapper[4750]: I0309 19:46:51.373829 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:46:51 crc kubenswrapper[4750]: E0309 19:46:51.374871 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:46:57 crc kubenswrapper[4750]: I0309 19:46:57.373602 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:46:57 crc kubenswrapper[4750]: E0309 19:46:57.374440 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:47:05 crc kubenswrapper[4750]: I0309 19:47:05.373798 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:47:05 crc kubenswrapper[4750]: E0309 19:47:05.374776 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:47:10 crc kubenswrapper[4750]: I0309 19:47:10.373847 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:47:10 crc kubenswrapper[4750]: E0309 19:47:10.374876 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:47:20 crc kubenswrapper[4750]: I0309 19:47:20.374291 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:47:20 crc kubenswrapper[4750]: E0309 19:47:20.375006 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:47:25 crc kubenswrapper[4750]: I0309 19:47:25.373513 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:47:25 crc kubenswrapper[4750]: E0309 19:47:25.375259 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:47:32 crc kubenswrapper[4750]: I0309 19:47:32.374001 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:47:32 crc kubenswrapper[4750]: I0309 19:47:32.808510 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"304f263c311bd521cb49fd14d8e5ef9f45fa6b8e8964420fac01b676716e348e"} Mar 09 19:47:39 crc kubenswrapper[4750]: I0309 19:47:39.385594 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:47:39 crc kubenswrapper[4750]: E0309 19:47:39.386319 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:47:53 crc kubenswrapper[4750]: I0309 19:47:53.373950 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:47:53 crc kubenswrapper[4750]: E0309 19:47:53.375035 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.182376 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551428-gmlh9"] Mar 09 19:48:00 crc kubenswrapper[4750]: E0309 19:48:00.183392 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="extract-content" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.183405 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="extract-content" Mar 09 19:48:00 crc kubenswrapper[4750]: E0309 19:48:00.183423 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="registry-server" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.183429 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="registry-server" Mar 09 19:48:00 crc kubenswrapper[4750]: E0309 19:48:00.183445 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="extract-utilities" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.183452 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="extract-utilities" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.183676 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="317e3659-3659-42d6-9b0d-1adc70754e5a" containerName="registry-server" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.184437 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551428-gmlh9" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.186607 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.186998 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.187179 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.209203 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551428-gmlh9"] Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.304139 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtf8z\" (UniqueName: \"kubernetes.io/projected/c85be130-e44d-4562-be4b-27f122d4ac14-kube-api-access-vtf8z\") pod \"auto-csr-approver-29551428-gmlh9\" (UID: \"c85be130-e44d-4562-be4b-27f122d4ac14\") " pod="openshift-infra/auto-csr-approver-29551428-gmlh9" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.406231 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtf8z\" (UniqueName: \"kubernetes.io/projected/c85be130-e44d-4562-be4b-27f122d4ac14-kube-api-access-vtf8z\") pod \"auto-csr-approver-29551428-gmlh9\" (UID: \"c85be130-e44d-4562-be4b-27f122d4ac14\") " pod="openshift-infra/auto-csr-approver-29551428-gmlh9" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.430536 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtf8z\" (UniqueName: \"kubernetes.io/projected/c85be130-e44d-4562-be4b-27f122d4ac14-kube-api-access-vtf8z\") pod \"auto-csr-approver-29551428-gmlh9\" (UID: \"c85be130-e44d-4562-be4b-27f122d4ac14\") " pod="openshift-infra/auto-csr-approver-29551428-gmlh9" Mar 09 19:48:00 crc kubenswrapper[4750]: I0309 19:48:00.513509 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551428-gmlh9" Mar 09 19:48:01 crc kubenswrapper[4750]: I0309 19:48:01.034841 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551428-gmlh9"] Mar 09 19:48:01 crc kubenswrapper[4750]: W0309 19:48:01.035562 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc85be130_e44d_4562_be4b_27f122d4ac14.slice/crio-643ae9c792b6d0cc00628aa72e6f9497743c4122e061b67c03a5c61571fcdc27 WatchSource:0}: Error finding container 643ae9c792b6d0cc00628aa72e6f9497743c4122e061b67c03a5c61571fcdc27: Status 404 returned error can't find the container with id 643ae9c792b6d0cc00628aa72e6f9497743c4122e061b67c03a5c61571fcdc27 Mar 09 19:48:01 crc kubenswrapper[4750]: I0309 19:48:01.168112 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551428-gmlh9" event={"ID":"c85be130-e44d-4562-be4b-27f122d4ac14","Type":"ContainerStarted","Data":"643ae9c792b6d0cc00628aa72e6f9497743c4122e061b67c03a5c61571fcdc27"} Mar 09 19:48:03 crc kubenswrapper[4750]: I0309 19:48:03.195467 4750 generic.go:334] "Generic (PLEG): container finished" podID="c85be130-e44d-4562-be4b-27f122d4ac14" containerID="7572658f4690237f90309914f8783a08389fb9a767f699bb3aa9cd10eecdd960" exitCode=0 Mar 09 19:48:03 crc kubenswrapper[4750]: I0309 19:48:03.195556 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551428-gmlh9" event={"ID":"c85be130-e44d-4562-be4b-27f122d4ac14","Type":"ContainerDied","Data":"7572658f4690237f90309914f8783a08389fb9a767f699bb3aa9cd10eecdd960"} Mar 09 19:48:05 crc kubenswrapper[4750]: I0309 19:48:05.213217 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551428-gmlh9" event={"ID":"c85be130-e44d-4562-be4b-27f122d4ac14","Type":"ContainerDied","Data":"643ae9c792b6d0cc00628aa72e6f9497743c4122e061b67c03a5c61571fcdc27"} Mar 09 19:48:05 crc kubenswrapper[4750]: I0309 19:48:05.213435 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="643ae9c792b6d0cc00628aa72e6f9497743c4122e061b67c03a5c61571fcdc27" Mar 09 19:48:05 crc kubenswrapper[4750]: I0309 19:48:05.316430 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551428-gmlh9" Mar 09 19:48:05 crc kubenswrapper[4750]: I0309 19:48:05.414736 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtf8z\" (UniqueName: \"kubernetes.io/projected/c85be130-e44d-4562-be4b-27f122d4ac14-kube-api-access-vtf8z\") pod \"c85be130-e44d-4562-be4b-27f122d4ac14\" (UID: \"c85be130-e44d-4562-be4b-27f122d4ac14\") " Mar 09 19:48:05 crc kubenswrapper[4750]: I0309 19:48:05.421150 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c85be130-e44d-4562-be4b-27f122d4ac14-kube-api-access-vtf8z" (OuterVolumeSpecName: "kube-api-access-vtf8z") pod "c85be130-e44d-4562-be4b-27f122d4ac14" (UID: "c85be130-e44d-4562-be4b-27f122d4ac14"). InnerVolumeSpecName "kube-api-access-vtf8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:48:05 crc kubenswrapper[4750]: I0309 19:48:05.517719 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtf8z\" (UniqueName: \"kubernetes.io/projected/c85be130-e44d-4562-be4b-27f122d4ac14-kube-api-access-vtf8z\") on node \"crc\" DevicePath \"\"" Mar 09 19:48:06 crc kubenswrapper[4750]: I0309 19:48:06.224683 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551428-gmlh9" Mar 09 19:48:06 crc kubenswrapper[4750]: I0309 19:48:06.394071 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551422-vv8tp"] Mar 09 19:48:06 crc kubenswrapper[4750]: I0309 19:48:06.405692 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551422-vv8tp"] Mar 09 19:48:07 crc kubenswrapper[4750]: I0309 19:48:07.402257 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9c60766-4c58-41b5-b11e-5b3f22578968" path="/var/lib/kubelet/pods/e9c60766-4c58-41b5-b11e-5b3f22578968/volumes" Mar 09 19:48:08 crc kubenswrapper[4750]: I0309 19:48:08.374013 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:48:08 crc kubenswrapper[4750]: E0309 19:48:08.375030 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:48:23 crc kubenswrapper[4750]: I0309 19:48:23.374923 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:48:23 crc kubenswrapper[4750]: E0309 19:48:23.376275 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:48:37 crc kubenswrapper[4750]: I0309 19:48:37.373834 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:48:37 crc kubenswrapper[4750]: E0309 19:48:37.375137 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:48:49 crc kubenswrapper[4750]: I0309 19:48:49.626724 4750 scope.go:117] "RemoveContainer" containerID="a7a390f650ca358270660b6eea88462a92367e7537d06fb527600069366bb7dc" Mar 09 19:48:50 crc kubenswrapper[4750]: I0309 19:48:50.373531 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:48:50 crc kubenswrapper[4750]: E0309 19:48:50.374202 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:49:02 crc kubenswrapper[4750]: I0309 19:49:02.373414 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:49:02 crc kubenswrapper[4750]: E0309 19:49:02.374527 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:49:16 crc kubenswrapper[4750]: I0309 19:49:16.373391 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:49:16 crc kubenswrapper[4750]: E0309 19:49:16.374319 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:49:29 crc kubenswrapper[4750]: I0309 19:49:29.390405 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:49:29 crc kubenswrapper[4750]: E0309 19:49:29.392024 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:49:44 crc kubenswrapper[4750]: I0309 19:49:44.373840 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:49:44 crc kubenswrapper[4750]: E0309 19:49:44.374641 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:49:51 crc kubenswrapper[4750]: I0309 19:49:51.742930 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:49:51 crc kubenswrapper[4750]: I0309 19:49:51.743650 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:49:55 crc kubenswrapper[4750]: I0309 19:49:55.374479 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:49:55 crc kubenswrapper[4750]: E0309 19:49:55.375403 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.170016 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551430-wnfl2"] Mar 09 19:50:00 crc kubenswrapper[4750]: E0309 19:50:00.171458 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c85be130-e44d-4562-be4b-27f122d4ac14" containerName="oc" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.171487 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c85be130-e44d-4562-be4b-27f122d4ac14" containerName="oc" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.171996 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c85be130-e44d-4562-be4b-27f122d4ac14" containerName="oc" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.173252 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551430-wnfl2" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.178560 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.178732 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.179202 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.183374 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551430-wnfl2"] Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.366220 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ctg7\" (UniqueName: \"kubernetes.io/projected/b150d127-1ec5-4ec6-8ff3-1541b355c39d-kube-api-access-7ctg7\") pod \"auto-csr-approver-29551430-wnfl2\" (UID: \"b150d127-1ec5-4ec6-8ff3-1541b355c39d\") " pod="openshift-infra/auto-csr-approver-29551430-wnfl2" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.468293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ctg7\" (UniqueName: \"kubernetes.io/projected/b150d127-1ec5-4ec6-8ff3-1541b355c39d-kube-api-access-7ctg7\") pod \"auto-csr-approver-29551430-wnfl2\" (UID: \"b150d127-1ec5-4ec6-8ff3-1541b355c39d\") " pod="openshift-infra/auto-csr-approver-29551430-wnfl2" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.492331 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ctg7\" (UniqueName: \"kubernetes.io/projected/b150d127-1ec5-4ec6-8ff3-1541b355c39d-kube-api-access-7ctg7\") pod \"auto-csr-approver-29551430-wnfl2\" (UID: \"b150d127-1ec5-4ec6-8ff3-1541b355c39d\") " pod="openshift-infra/auto-csr-approver-29551430-wnfl2" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.504006 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551430-wnfl2" Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.985835 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551430-wnfl2"] Mar 09 19:50:00 crc kubenswrapper[4750]: I0309 19:50:00.986118 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:50:01 crc kubenswrapper[4750]: I0309 19:50:01.595854 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551430-wnfl2" event={"ID":"b150d127-1ec5-4ec6-8ff3-1541b355c39d","Type":"ContainerStarted","Data":"1e0f22fe6063a613e6763fe873ad8df5f964b6081c154dddce6ac2fff3cf192c"} Mar 09 19:50:03 crc kubenswrapper[4750]: I0309 19:50:03.617257 4750 generic.go:334] "Generic (PLEG): container finished" podID="b150d127-1ec5-4ec6-8ff3-1541b355c39d" containerID="0ebcaccb501c5f98438b7b26063abee356b333cfe18196f1f25962a62ef03a36" exitCode=0 Mar 09 19:50:03 crc kubenswrapper[4750]: I0309 19:50:03.617368 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551430-wnfl2" event={"ID":"b150d127-1ec5-4ec6-8ff3-1541b355c39d","Type":"ContainerDied","Data":"0ebcaccb501c5f98438b7b26063abee356b333cfe18196f1f25962a62ef03a36"} Mar 09 19:50:04 crc kubenswrapper[4750]: I0309 19:50:04.995127 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551430-wnfl2" Mar 09 19:50:05 crc kubenswrapper[4750]: I0309 19:50:05.172414 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ctg7\" (UniqueName: \"kubernetes.io/projected/b150d127-1ec5-4ec6-8ff3-1541b355c39d-kube-api-access-7ctg7\") pod \"b150d127-1ec5-4ec6-8ff3-1541b355c39d\" (UID: \"b150d127-1ec5-4ec6-8ff3-1541b355c39d\") " Mar 09 19:50:05 crc kubenswrapper[4750]: I0309 19:50:05.179067 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b150d127-1ec5-4ec6-8ff3-1541b355c39d-kube-api-access-7ctg7" (OuterVolumeSpecName: "kube-api-access-7ctg7") pod "b150d127-1ec5-4ec6-8ff3-1541b355c39d" (UID: "b150d127-1ec5-4ec6-8ff3-1541b355c39d"). InnerVolumeSpecName "kube-api-access-7ctg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:50:05 crc kubenswrapper[4750]: I0309 19:50:05.275960 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ctg7\" (UniqueName: \"kubernetes.io/projected/b150d127-1ec5-4ec6-8ff3-1541b355c39d-kube-api-access-7ctg7\") on node \"crc\" DevicePath \"\"" Mar 09 19:50:05 crc kubenswrapper[4750]: I0309 19:50:05.636918 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551430-wnfl2" event={"ID":"b150d127-1ec5-4ec6-8ff3-1541b355c39d","Type":"ContainerDied","Data":"1e0f22fe6063a613e6763fe873ad8df5f964b6081c154dddce6ac2fff3cf192c"} Mar 09 19:50:05 crc kubenswrapper[4750]: I0309 19:50:05.637042 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0f22fe6063a613e6763fe873ad8df5f964b6081c154dddce6ac2fff3cf192c" Mar 09 19:50:05 crc kubenswrapper[4750]: I0309 19:50:05.636971 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551430-wnfl2" Mar 09 19:50:06 crc kubenswrapper[4750]: I0309 19:50:06.084288 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551424-bnw9n"] Mar 09 19:50:06 crc kubenswrapper[4750]: I0309 19:50:06.094483 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551424-bnw9n"] Mar 09 19:50:07 crc kubenswrapper[4750]: I0309 19:50:07.373794 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:50:07 crc kubenswrapper[4750]: E0309 19:50:07.374114 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:50:07 crc kubenswrapper[4750]: I0309 19:50:07.388372 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0822dfea-565b-4508-a995-83a244b0c492" path="/var/lib/kubelet/pods/0822dfea-565b-4508-a995-83a244b0c492/volumes" Mar 09 19:50:20 crc kubenswrapper[4750]: I0309 19:50:20.373353 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:50:20 crc kubenswrapper[4750]: E0309 19:50:20.374412 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:50:21 crc kubenswrapper[4750]: I0309 19:50:21.743810 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:50:21 crc kubenswrapper[4750]: I0309 19:50:21.744171 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:50:33 crc kubenswrapper[4750]: I0309 19:50:33.373952 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:50:33 crc kubenswrapper[4750]: E0309 19:50:33.374945 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:50:47 crc kubenswrapper[4750]: I0309 19:50:47.373820 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:50:49 crc kubenswrapper[4750]: I0309 19:50:49.211968 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d"} Mar 09 19:50:49 crc kubenswrapper[4750]: I0309 19:50:49.761598 4750 scope.go:117] "RemoveContainer" containerID="a920df654f36f88feb468391f36876124b6942f5bd33e934e944a3903e68a669" Mar 09 19:50:50 crc kubenswrapper[4750]: I0309 19:50:50.189860 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:50:51 crc kubenswrapper[4750]: I0309 19:50:51.743234 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:50:51 crc kubenswrapper[4750]: I0309 19:50:51.743696 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:50:51 crc kubenswrapper[4750]: I0309 19:50:51.743765 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:50:51 crc kubenswrapper[4750]: I0309 19:50:51.745030 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"304f263c311bd521cb49fd14d8e5ef9f45fa6b8e8964420fac01b676716e348e"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:50:51 crc kubenswrapper[4750]: I0309 19:50:51.745134 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://304f263c311bd521cb49fd14d8e5ef9f45fa6b8e8964420fac01b676716e348e" gracePeriod=600 Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.258447 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" exitCode=1 Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.258542 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d"} Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.259040 4750 scope.go:117] "RemoveContainer" containerID="10b27cb60244ab944c60e99a8e56c264f769c37c9450c317eacea9e3928a1188" Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.260541 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:50:52 crc kubenswrapper[4750]: E0309 19:50:52.261681 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.265703 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="304f263c311bd521cb49fd14d8e5ef9f45fa6b8e8964420fac01b676716e348e" exitCode=0 Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.265757 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"304f263c311bd521cb49fd14d8e5ef9f45fa6b8e8964420fac01b676716e348e"} Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.265793 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768"} Mar 09 19:50:52 crc kubenswrapper[4750]: I0309 19:50:52.334391 4750 scope.go:117] "RemoveContainer" containerID="991eebb014df52ead95e915299b035a03f3f181d8db37eb742016e6212e8ffbc" Mar 09 19:50:53 crc kubenswrapper[4750]: I0309 19:50:53.189973 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:50:53 crc kubenswrapper[4750]: I0309 19:50:53.337328 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:50:53 crc kubenswrapper[4750]: E0309 19:50:53.338044 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:50:55 crc kubenswrapper[4750]: I0309 19:50:55.190385 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:50:55 crc kubenswrapper[4750]: I0309 19:50:55.192082 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:50:55 crc kubenswrapper[4750]: E0309 19:50:55.192495 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:51:10 crc kubenswrapper[4750]: I0309 19:51:10.374559 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:51:10 crc kubenswrapper[4750]: E0309 19:51:10.375699 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:51:22 crc kubenswrapper[4750]: I0309 19:51:22.373065 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:51:22 crc kubenswrapper[4750]: E0309 19:51:22.373844 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:51:35 crc kubenswrapper[4750]: I0309 19:51:35.373492 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:51:35 crc kubenswrapper[4750]: E0309 19:51:35.374051 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:51:48 crc kubenswrapper[4750]: I0309 19:51:48.373268 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:51:48 crc kubenswrapper[4750]: E0309 19:51:48.373982 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:51:59 crc kubenswrapper[4750]: I0309 19:51:59.384436 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:51:59 crc kubenswrapper[4750]: E0309 19:51:59.387707 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.167764 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551432-2mjmk"] Mar 09 19:52:00 crc kubenswrapper[4750]: E0309 19:52:00.168673 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b150d127-1ec5-4ec6-8ff3-1541b355c39d" containerName="oc" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.168696 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b150d127-1ec5-4ec6-8ff3-1541b355c39d" containerName="oc" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.168896 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b150d127-1ec5-4ec6-8ff3-1541b355c39d" containerName="oc" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.169611 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551432-2mjmk" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.173146 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.173489 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.175909 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.202240 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551432-2mjmk"] Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.247092 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmvvm\" (UniqueName: \"kubernetes.io/projected/f5262a0c-eea0-42c8-9b41-20d2045e593d-kube-api-access-kmvvm\") pod \"auto-csr-approver-29551432-2mjmk\" (UID: \"f5262a0c-eea0-42c8-9b41-20d2045e593d\") " pod="openshift-infra/auto-csr-approver-29551432-2mjmk" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.349246 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmvvm\" (UniqueName: \"kubernetes.io/projected/f5262a0c-eea0-42c8-9b41-20d2045e593d-kube-api-access-kmvvm\") pod \"auto-csr-approver-29551432-2mjmk\" (UID: \"f5262a0c-eea0-42c8-9b41-20d2045e593d\") " pod="openshift-infra/auto-csr-approver-29551432-2mjmk" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.380602 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmvvm\" (UniqueName: \"kubernetes.io/projected/f5262a0c-eea0-42c8-9b41-20d2045e593d-kube-api-access-kmvvm\") pod \"auto-csr-approver-29551432-2mjmk\" (UID: \"f5262a0c-eea0-42c8-9b41-20d2045e593d\") " pod="openshift-infra/auto-csr-approver-29551432-2mjmk" Mar 09 19:52:00 crc kubenswrapper[4750]: I0309 19:52:00.505617 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551432-2mjmk" Mar 09 19:52:01 crc kubenswrapper[4750]: I0309 19:52:01.614444 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551432-2mjmk"] Mar 09 19:52:02 crc kubenswrapper[4750]: I0309 19:52:02.182970 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551432-2mjmk" event={"ID":"f5262a0c-eea0-42c8-9b41-20d2045e593d","Type":"ContainerStarted","Data":"c6e667f7c1365e3c992309ad7a8c295eee393aaaa84ea596593984d44d5e529d"} Mar 09 19:52:03 crc kubenswrapper[4750]: I0309 19:52:03.196722 4750 generic.go:334] "Generic (PLEG): container finished" podID="f5262a0c-eea0-42c8-9b41-20d2045e593d" containerID="78e18030fadfc0ef37e45f340a312c160da6c8e354be867e652389be5bfc4fe1" exitCode=0 Mar 09 19:52:03 crc kubenswrapper[4750]: I0309 19:52:03.196795 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551432-2mjmk" event={"ID":"f5262a0c-eea0-42c8-9b41-20d2045e593d","Type":"ContainerDied","Data":"78e18030fadfc0ef37e45f340a312c160da6c8e354be867e652389be5bfc4fe1"} Mar 09 19:52:04 crc kubenswrapper[4750]: I0309 19:52:04.587946 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551432-2mjmk" Mar 09 19:52:04 crc kubenswrapper[4750]: I0309 19:52:04.651297 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmvvm\" (UniqueName: \"kubernetes.io/projected/f5262a0c-eea0-42c8-9b41-20d2045e593d-kube-api-access-kmvvm\") pod \"f5262a0c-eea0-42c8-9b41-20d2045e593d\" (UID: \"f5262a0c-eea0-42c8-9b41-20d2045e593d\") " Mar 09 19:52:04 crc kubenswrapper[4750]: I0309 19:52:04.656511 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5262a0c-eea0-42c8-9b41-20d2045e593d-kube-api-access-kmvvm" (OuterVolumeSpecName: "kube-api-access-kmvvm") pod "f5262a0c-eea0-42c8-9b41-20d2045e593d" (UID: "f5262a0c-eea0-42c8-9b41-20d2045e593d"). InnerVolumeSpecName "kube-api-access-kmvvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:52:04 crc kubenswrapper[4750]: I0309 19:52:04.753723 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmvvm\" (UniqueName: \"kubernetes.io/projected/f5262a0c-eea0-42c8-9b41-20d2045e593d-kube-api-access-kmvvm\") on node \"crc\" DevicePath \"\"" Mar 09 19:52:05 crc kubenswrapper[4750]: I0309 19:52:05.215493 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551432-2mjmk" event={"ID":"f5262a0c-eea0-42c8-9b41-20d2045e593d","Type":"ContainerDied","Data":"c6e667f7c1365e3c992309ad7a8c295eee393aaaa84ea596593984d44d5e529d"} Mar 09 19:52:05 crc kubenswrapper[4750]: I0309 19:52:05.215710 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6e667f7c1365e3c992309ad7a8c295eee393aaaa84ea596593984d44d5e529d" Mar 09 19:52:05 crc kubenswrapper[4750]: I0309 19:52:05.215596 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551432-2mjmk" Mar 09 19:52:05 crc kubenswrapper[4750]: I0309 19:52:05.729420 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551426-v2pm4"] Mar 09 19:52:05 crc kubenswrapper[4750]: I0309 19:52:05.747345 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551426-v2pm4"] Mar 09 19:52:07 crc kubenswrapper[4750]: I0309 19:52:07.386291 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4948b4b-7c80-4298-913b-2d741c6b7a84" path="/var/lib/kubelet/pods/b4948b4b-7c80-4298-913b-2d741c6b7a84/volumes" Mar 09 19:52:14 crc kubenswrapper[4750]: I0309 19:52:14.373336 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:52:14 crc kubenswrapper[4750]: E0309 19:52:14.374292 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:52:26 crc kubenswrapper[4750]: I0309 19:52:26.374501 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:52:26 crc kubenswrapper[4750]: E0309 19:52:26.376071 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:52:40 crc kubenswrapper[4750]: I0309 19:52:40.373426 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:52:40 crc kubenswrapper[4750]: E0309 19:52:40.375899 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:52:49 crc kubenswrapper[4750]: I0309 19:52:49.902190 4750 scope.go:117] "RemoveContainer" containerID="87f8f9255d585d2fc0a288a3e14089e5379e67edf11374f9f039c239a58288f7" Mar 09 19:52:52 crc kubenswrapper[4750]: I0309 19:52:52.375920 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:52:52 crc kubenswrapper[4750]: E0309 19:52:52.376972 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:53:05 crc kubenswrapper[4750]: I0309 19:53:05.373927 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:53:05 crc kubenswrapper[4750]: E0309 19:53:05.374986 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:53:16 crc kubenswrapper[4750]: I0309 19:53:16.376536 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:53:16 crc kubenswrapper[4750]: E0309 19:53:16.377896 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:53:21 crc kubenswrapper[4750]: I0309 19:53:21.743225 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:53:21 crc kubenswrapper[4750]: I0309 19:53:21.743867 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:53:28 crc kubenswrapper[4750]: I0309 19:53:28.373879 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:53:28 crc kubenswrapper[4750]: E0309 19:53:28.375225 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:53:43 crc kubenswrapper[4750]: I0309 19:53:43.373646 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:53:43 crc kubenswrapper[4750]: E0309 19:53:43.374941 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:53:51 crc kubenswrapper[4750]: I0309 19:53:51.743250 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:53:51 crc kubenswrapper[4750]: I0309 19:53:51.743986 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:53:55 crc kubenswrapper[4750]: I0309 19:53:55.373694 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:53:55 crc kubenswrapper[4750]: E0309 19:53:55.375249 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.169409 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551434-2gxz9"] Mar 09 19:54:00 crc kubenswrapper[4750]: E0309 19:54:00.170364 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5262a0c-eea0-42c8-9b41-20d2045e593d" containerName="oc" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.170379 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5262a0c-eea0-42c8-9b41-20d2045e593d" containerName="oc" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.170567 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5262a0c-eea0-42c8-9b41-20d2045e593d" containerName="oc" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.171252 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551434-2gxz9" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.179048 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.179493 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.187320 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551434-2gxz9"] Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.187498 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.327433 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbvkk\" (UniqueName: \"kubernetes.io/projected/e14dabf4-c85c-481b-8025-d4874c8b1a8a-kube-api-access-vbvkk\") pod \"auto-csr-approver-29551434-2gxz9\" (UID: \"e14dabf4-c85c-481b-8025-d4874c8b1a8a\") " pod="openshift-infra/auto-csr-approver-29551434-2gxz9" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.430292 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbvkk\" (UniqueName: \"kubernetes.io/projected/e14dabf4-c85c-481b-8025-d4874c8b1a8a-kube-api-access-vbvkk\") pod \"auto-csr-approver-29551434-2gxz9\" (UID: \"e14dabf4-c85c-481b-8025-d4874c8b1a8a\") " pod="openshift-infra/auto-csr-approver-29551434-2gxz9" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.457607 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbvkk\" (UniqueName: \"kubernetes.io/projected/e14dabf4-c85c-481b-8025-d4874c8b1a8a-kube-api-access-vbvkk\") pod \"auto-csr-approver-29551434-2gxz9\" (UID: \"e14dabf4-c85c-481b-8025-d4874c8b1a8a\") " pod="openshift-infra/auto-csr-approver-29551434-2gxz9" Mar 09 19:54:00 crc kubenswrapper[4750]: I0309 19:54:00.526971 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551434-2gxz9" Mar 09 19:54:01 crc kubenswrapper[4750]: I0309 19:54:01.035615 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551434-2gxz9"] Mar 09 19:54:01 crc kubenswrapper[4750]: I0309 19:54:01.228139 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551434-2gxz9" event={"ID":"e14dabf4-c85c-481b-8025-d4874c8b1a8a","Type":"ContainerStarted","Data":"104ab48d736eab37313aade9d71a7c8f3763c5a560058d72e0fcc3b892520ff2"} Mar 09 19:54:03 crc kubenswrapper[4750]: I0309 19:54:03.252502 4750 generic.go:334] "Generic (PLEG): container finished" podID="e14dabf4-c85c-481b-8025-d4874c8b1a8a" containerID="bbbb6ec93f7b95cfa24c9cfd98b584ddb549fcd9be22f020b40929d30b2cdca0" exitCode=0 Mar 09 19:54:03 crc kubenswrapper[4750]: I0309 19:54:03.252569 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551434-2gxz9" event={"ID":"e14dabf4-c85c-481b-8025-d4874c8b1a8a","Type":"ContainerDied","Data":"bbbb6ec93f7b95cfa24c9cfd98b584ddb549fcd9be22f020b40929d30b2cdca0"} Mar 09 19:54:04 crc kubenswrapper[4750]: I0309 19:54:04.699743 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551434-2gxz9" Mar 09 19:54:04 crc kubenswrapper[4750]: I0309 19:54:04.832811 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbvkk\" (UniqueName: \"kubernetes.io/projected/e14dabf4-c85c-481b-8025-d4874c8b1a8a-kube-api-access-vbvkk\") pod \"e14dabf4-c85c-481b-8025-d4874c8b1a8a\" (UID: \"e14dabf4-c85c-481b-8025-d4874c8b1a8a\") " Mar 09 19:54:04 crc kubenswrapper[4750]: I0309 19:54:04.845135 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e14dabf4-c85c-481b-8025-d4874c8b1a8a-kube-api-access-vbvkk" (OuterVolumeSpecName: "kube-api-access-vbvkk") pod "e14dabf4-c85c-481b-8025-d4874c8b1a8a" (UID: "e14dabf4-c85c-481b-8025-d4874c8b1a8a"). InnerVolumeSpecName "kube-api-access-vbvkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:54:04 crc kubenswrapper[4750]: I0309 19:54:04.936294 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbvkk\" (UniqueName: \"kubernetes.io/projected/e14dabf4-c85c-481b-8025-d4874c8b1a8a-kube-api-access-vbvkk\") on node \"crc\" DevicePath \"\"" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.254547 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bbgcr"] Mar 09 19:54:05 crc kubenswrapper[4750]: E0309 19:54:05.255082 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14dabf4-c85c-481b-8025-d4874c8b1a8a" containerName="oc" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.255102 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14dabf4-c85c-481b-8025-d4874c8b1a8a" containerName="oc" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.255387 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e14dabf4-c85c-481b-8025-d4874c8b1a8a" containerName="oc" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.257380 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.275440 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbgcr"] Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.330708 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551434-2gxz9" event={"ID":"e14dabf4-c85c-481b-8025-d4874c8b1a8a","Type":"ContainerDied","Data":"104ab48d736eab37313aade9d71a7c8f3763c5a560058d72e0fcc3b892520ff2"} Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.330773 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="104ab48d736eab37313aade9d71a7c8f3763c5a560058d72e0fcc3b892520ff2" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.330828 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551434-2gxz9" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.346211 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-catalog-content\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.346301 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z5d4\" (UniqueName: \"kubernetes.io/projected/00890e38-69bf-44ea-9c90-4bca46561741-kube-api-access-5z5d4\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.346624 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-utilities\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.448254 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z5d4\" (UniqueName: \"kubernetes.io/projected/00890e38-69bf-44ea-9c90-4bca46561741-kube-api-access-5z5d4\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.448412 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-utilities\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.448884 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-utilities\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.449186 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-catalog-content\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.448972 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-catalog-content\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.467989 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z5d4\" (UniqueName: \"kubernetes.io/projected/00890e38-69bf-44ea-9c90-4bca46561741-kube-api-access-5z5d4\") pod \"redhat-operators-bbgcr\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.633986 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.780619 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551428-gmlh9"] Mar 09 19:54:05 crc kubenswrapper[4750]: I0309 19:54:05.790800 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551428-gmlh9"] Mar 09 19:54:06 crc kubenswrapper[4750]: I0309 19:54:06.119460 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbgcr"] Mar 09 19:54:06 crc kubenswrapper[4750]: W0309 19:54:06.120300 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00890e38_69bf_44ea_9c90_4bca46561741.slice/crio-ac3f465a5605a7307232185538994bf7d8073b094910aac274e07525df6a13fc WatchSource:0}: Error finding container ac3f465a5605a7307232185538994bf7d8073b094910aac274e07525df6a13fc: Status 404 returned error can't find the container with id ac3f465a5605a7307232185538994bf7d8073b094910aac274e07525df6a13fc Mar 09 19:54:06 crc kubenswrapper[4750]: I0309 19:54:06.340997 4750 generic.go:334] "Generic (PLEG): container finished" podID="00890e38-69bf-44ea-9c90-4bca46561741" containerID="7838a59444f0d8d58c3612f05b8fc5e01b920b063da0f726328ec6bc624aa173" exitCode=0 Mar 09 19:54:06 crc kubenswrapper[4750]: I0309 19:54:06.341110 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbgcr" event={"ID":"00890e38-69bf-44ea-9c90-4bca46561741","Type":"ContainerDied","Data":"7838a59444f0d8d58c3612f05b8fc5e01b920b063da0f726328ec6bc624aa173"} Mar 09 19:54:06 crc kubenswrapper[4750]: I0309 19:54:06.341308 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbgcr" event={"ID":"00890e38-69bf-44ea-9c90-4bca46561741","Type":"ContainerStarted","Data":"ac3f465a5605a7307232185538994bf7d8073b094910aac274e07525df6a13fc"} Mar 09 19:54:07 crc kubenswrapper[4750]: I0309 19:54:07.351122 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbgcr" event={"ID":"00890e38-69bf-44ea-9c90-4bca46561741","Type":"ContainerStarted","Data":"9803ec38e2477ee94c5282a9e4eb36a747f9ff901fd7b68a533a53c7e9f1f7b5"} Mar 09 19:54:07 crc kubenswrapper[4750]: I0309 19:54:07.385239 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c85be130-e44d-4562-be4b-27f122d4ac14" path="/var/lib/kubelet/pods/c85be130-e44d-4562-be4b-27f122d4ac14/volumes" Mar 09 19:54:10 crc kubenswrapper[4750]: I0309 19:54:10.398336 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:54:10 crc kubenswrapper[4750]: E0309 19:54:10.399915 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:54:12 crc kubenswrapper[4750]: I0309 19:54:12.437214 4750 generic.go:334] "Generic (PLEG): container finished" podID="00890e38-69bf-44ea-9c90-4bca46561741" containerID="9803ec38e2477ee94c5282a9e4eb36a747f9ff901fd7b68a533a53c7e9f1f7b5" exitCode=0 Mar 09 19:54:12 crc kubenswrapper[4750]: I0309 19:54:12.438305 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbgcr" event={"ID":"00890e38-69bf-44ea-9c90-4bca46561741","Type":"ContainerDied","Data":"9803ec38e2477ee94c5282a9e4eb36a747f9ff901fd7b68a533a53c7e9f1f7b5"} Mar 09 19:54:13 crc kubenswrapper[4750]: I0309 19:54:13.450562 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbgcr" event={"ID":"00890e38-69bf-44ea-9c90-4bca46561741","Type":"ContainerStarted","Data":"3680ca273c2c4465f9cae1e9c7187c9edfdaa651a28f13050068f97698f864ca"} Mar 09 19:54:13 crc kubenswrapper[4750]: I0309 19:54:13.490558 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bbgcr" podStartSLOduration=1.870859214 podStartE2EDuration="8.490527821s" podCreationTimestamp="2026-03-09 19:54:05 +0000 UTC" firstStartedPulling="2026-03-09 19:54:06.342499362 +0000 UTC m=+5327.684971770" lastFinishedPulling="2026-03-09 19:54:12.962167969 +0000 UTC m=+5334.304640377" observedRunningTime="2026-03-09 19:54:13.475873046 +0000 UTC m=+5334.818345454" watchObservedRunningTime="2026-03-09 19:54:13.490527821 +0000 UTC m=+5334.833000249" Mar 09 19:54:15 crc kubenswrapper[4750]: I0309 19:54:15.634944 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:15 crc kubenswrapper[4750]: I0309 19:54:15.635522 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:16 crc kubenswrapper[4750]: I0309 19:54:16.711936 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bbgcr" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="registry-server" probeResult="failure" output=< Mar 09 19:54:16 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 19:54:16 crc kubenswrapper[4750]: > Mar 09 19:54:21 crc kubenswrapper[4750]: I0309 19:54:21.743654 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 19:54:21 crc kubenswrapper[4750]: I0309 19:54:21.744181 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 19:54:21 crc kubenswrapper[4750]: I0309 19:54:21.744232 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 19:54:21 crc kubenswrapper[4750]: I0309 19:54:21.744940 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 19:54:21 crc kubenswrapper[4750]: I0309 19:54:21.745007 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" gracePeriod=600 Mar 09 19:54:21 crc kubenswrapper[4750]: E0309 19:54:21.883784 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:54:22 crc kubenswrapper[4750]: I0309 19:54:22.374954 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:54:22 crc kubenswrapper[4750]: E0309 19:54:22.375416 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:54:22 crc kubenswrapper[4750]: I0309 19:54:22.567269 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" exitCode=0 Mar 09 19:54:22 crc kubenswrapper[4750]: I0309 19:54:22.567331 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768"} Mar 09 19:54:22 crc kubenswrapper[4750]: I0309 19:54:22.567421 4750 scope.go:117] "RemoveContainer" containerID="304f263c311bd521cb49fd14d8e5ef9f45fa6b8e8964420fac01b676716e348e" Mar 09 19:54:22 crc kubenswrapper[4750]: I0309 19:54:22.568402 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:54:22 crc kubenswrapper[4750]: E0309 19:54:22.569111 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:54:25 crc kubenswrapper[4750]: I0309 19:54:25.710361 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:25 crc kubenswrapper[4750]: I0309 19:54:25.768615 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:25 crc kubenswrapper[4750]: I0309 19:54:25.960587 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbgcr"] Mar 09 19:54:27 crc kubenswrapper[4750]: I0309 19:54:27.639029 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bbgcr" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="registry-server" containerID="cri-o://3680ca273c2c4465f9cae1e9c7187c9edfdaa651a28f13050068f97698f864ca" gracePeriod=2 Mar 09 19:54:28 crc kubenswrapper[4750]: I0309 19:54:28.656966 4750 generic.go:334] "Generic (PLEG): container finished" podID="00890e38-69bf-44ea-9c90-4bca46561741" containerID="3680ca273c2c4465f9cae1e9c7187c9edfdaa651a28f13050068f97698f864ca" exitCode=0 Mar 09 19:54:28 crc kubenswrapper[4750]: I0309 19:54:28.657024 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbgcr" event={"ID":"00890e38-69bf-44ea-9c90-4bca46561741","Type":"ContainerDied","Data":"3680ca273c2c4465f9cae1e9c7187c9edfdaa651a28f13050068f97698f864ca"} Mar 09 19:54:28 crc kubenswrapper[4750]: I0309 19:54:28.657413 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbgcr" event={"ID":"00890e38-69bf-44ea-9c90-4bca46561741","Type":"ContainerDied","Data":"ac3f465a5605a7307232185538994bf7d8073b094910aac274e07525df6a13fc"} Mar 09 19:54:28 crc kubenswrapper[4750]: I0309 19:54:28.657440 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac3f465a5605a7307232185538994bf7d8073b094910aac274e07525df6a13fc" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.452429 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.590951 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-utilities\") pod \"00890e38-69bf-44ea-9c90-4bca46561741\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.591041 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-catalog-content\") pod \"00890e38-69bf-44ea-9c90-4bca46561741\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.591194 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z5d4\" (UniqueName: \"kubernetes.io/projected/00890e38-69bf-44ea-9c90-4bca46561741-kube-api-access-5z5d4\") pod \"00890e38-69bf-44ea-9c90-4bca46561741\" (UID: \"00890e38-69bf-44ea-9c90-4bca46561741\") " Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.591612 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-utilities" (OuterVolumeSpecName: "utilities") pod "00890e38-69bf-44ea-9c90-4bca46561741" (UID: "00890e38-69bf-44ea-9c90-4bca46561741"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.591774 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.599352 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00890e38-69bf-44ea-9c90-4bca46561741-kube-api-access-5z5d4" (OuterVolumeSpecName: "kube-api-access-5z5d4") pod "00890e38-69bf-44ea-9c90-4bca46561741" (UID: "00890e38-69bf-44ea-9c90-4bca46561741"). InnerVolumeSpecName "kube-api-access-5z5d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.671602 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbgcr" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.693537 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z5d4\" (UniqueName: \"kubernetes.io/projected/00890e38-69bf-44ea-9c90-4bca46561741-kube-api-access-5z5d4\") on node \"crc\" DevicePath \"\"" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.730516 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00890e38-69bf-44ea-9c90-4bca46561741" (UID: "00890e38-69bf-44ea-9c90-4bca46561741"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:54:29 crc kubenswrapper[4750]: I0309 19:54:29.795491 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00890e38-69bf-44ea-9c90-4bca46561741-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:54:30 crc kubenswrapper[4750]: I0309 19:54:30.030455 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbgcr"] Mar 09 19:54:30 crc kubenswrapper[4750]: I0309 19:54:30.047257 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bbgcr"] Mar 09 19:54:31 crc kubenswrapper[4750]: I0309 19:54:31.405269 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00890e38-69bf-44ea-9c90-4bca46561741" path="/var/lib/kubelet/pods/00890e38-69bf-44ea-9c90-4bca46561741/volumes" Mar 09 19:54:35 crc kubenswrapper[4750]: I0309 19:54:35.375209 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:54:35 crc kubenswrapper[4750]: E0309 19:54:35.376575 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:54:36 crc kubenswrapper[4750]: I0309 19:54:36.375458 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:54:36 crc kubenswrapper[4750]: E0309 19:54:36.376278 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:54:48 crc kubenswrapper[4750]: I0309 19:54:48.374565 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:54:48 crc kubenswrapper[4750]: E0309 19:54:48.375449 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:54:49 crc kubenswrapper[4750]: I0309 19:54:49.380526 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:54:49 crc kubenswrapper[4750]: E0309 19:54:49.381075 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:54:50 crc kubenswrapper[4750]: I0309 19:54:50.007255 4750 scope.go:117] "RemoveContainer" containerID="7572658f4690237f90309914f8783a08389fb9a767f699bb3aa9cd10eecdd960" Mar 09 19:54:59 crc kubenswrapper[4750]: I0309 19:54:59.382722 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:54:59 crc kubenswrapper[4750]: E0309 19:54:59.383851 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:55:01 crc kubenswrapper[4750]: I0309 19:55:01.374458 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:55:01 crc kubenswrapper[4750]: E0309 19:55:01.374772 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:55:11 crc kubenswrapper[4750]: I0309 19:55:11.374425 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:55:11 crc kubenswrapper[4750]: E0309 19:55:11.375839 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:55:16 crc kubenswrapper[4750]: I0309 19:55:16.374727 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:55:16 crc kubenswrapper[4750]: E0309 19:55:16.375875 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:55:24 crc kubenswrapper[4750]: I0309 19:55:24.374141 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:55:24 crc kubenswrapper[4750]: E0309 19:55:24.375093 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:55:28 crc kubenswrapper[4750]: I0309 19:55:28.374559 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:55:28 crc kubenswrapper[4750]: E0309 19:55:28.375804 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:55:38 crc kubenswrapper[4750]: I0309 19:55:38.373227 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:55:38 crc kubenswrapper[4750]: E0309 19:55:38.374023 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:55:39 crc kubenswrapper[4750]: I0309 19:55:39.384894 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:55:39 crc kubenswrapper[4750]: E0309 19:55:39.385624 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:55:49 crc kubenswrapper[4750]: I0309 19:55:49.400698 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:55:49 crc kubenswrapper[4750]: E0309 19:55:49.401758 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:55:50 crc kubenswrapper[4750]: I0309 19:55:50.374150 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:55:50 crc kubenswrapper[4750]: E0309 19:55:50.374819 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.141103 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551436-s4fgg"] Mar 09 19:56:00 crc kubenswrapper[4750]: E0309 19:56:00.141912 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="extract-utilities" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.141923 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="extract-utilities" Mar 09 19:56:00 crc kubenswrapper[4750]: E0309 19:56:00.141950 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="extract-content" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.141956 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="extract-content" Mar 09 19:56:00 crc kubenswrapper[4750]: E0309 19:56:00.141978 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="registry-server" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.141983 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="registry-server" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.142464 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="00890e38-69bf-44ea-9c90-4bca46561741" containerName="registry-server" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.143170 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551436-s4fgg" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.145403 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.145555 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.145663 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.157229 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551436-s4fgg"] Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.265375 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxnpc\" (UniqueName: \"kubernetes.io/projected/754c550f-6079-4f6d-8f18-af4ebddfdf61-kube-api-access-wxnpc\") pod \"auto-csr-approver-29551436-s4fgg\" (UID: \"754c550f-6079-4f6d-8f18-af4ebddfdf61\") " pod="openshift-infra/auto-csr-approver-29551436-s4fgg" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.367788 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxnpc\" (UniqueName: \"kubernetes.io/projected/754c550f-6079-4f6d-8f18-af4ebddfdf61-kube-api-access-wxnpc\") pod \"auto-csr-approver-29551436-s4fgg\" (UID: \"754c550f-6079-4f6d-8f18-af4ebddfdf61\") " pod="openshift-infra/auto-csr-approver-29551436-s4fgg" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.397059 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxnpc\" (UniqueName: \"kubernetes.io/projected/754c550f-6079-4f6d-8f18-af4ebddfdf61-kube-api-access-wxnpc\") pod \"auto-csr-approver-29551436-s4fgg\" (UID: \"754c550f-6079-4f6d-8f18-af4ebddfdf61\") " pod="openshift-infra/auto-csr-approver-29551436-s4fgg" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.461668 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551436-s4fgg" Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.979316 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551436-s4fgg"] Mar 09 19:56:00 crc kubenswrapper[4750]: W0309 19:56:00.989515 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod754c550f_6079_4f6d_8f18_af4ebddfdf61.slice/crio-36a59350874f2afae09ffd739490661c9cf13be0297bbe714b4669db2340acd2 WatchSource:0}: Error finding container 36a59350874f2afae09ffd739490661c9cf13be0297bbe714b4669db2340acd2: Status 404 returned error can't find the container with id 36a59350874f2afae09ffd739490661c9cf13be0297bbe714b4669db2340acd2 Mar 09 19:56:00 crc kubenswrapper[4750]: I0309 19:56:00.992447 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 19:56:01 crc kubenswrapper[4750]: I0309 19:56:01.792164 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551436-s4fgg" event={"ID":"754c550f-6079-4f6d-8f18-af4ebddfdf61","Type":"ContainerStarted","Data":"36a59350874f2afae09ffd739490661c9cf13be0297bbe714b4669db2340acd2"} Mar 09 19:56:02 crc kubenswrapper[4750]: I0309 19:56:02.373559 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:56:02 crc kubenswrapper[4750]: E0309 19:56:02.374253 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:56:03 crc kubenswrapper[4750]: I0309 19:56:03.373483 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:56:03 crc kubenswrapper[4750]: I0309 19:56:03.838172 4750 generic.go:334] "Generic (PLEG): container finished" podID="754c550f-6079-4f6d-8f18-af4ebddfdf61" containerID="aec92e10d4cb243e09ecb9e45935ad305c576ca83a2923a10cd6ad13d78de63b" exitCode=0 Mar 09 19:56:03 crc kubenswrapper[4750]: I0309 19:56:03.838232 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551436-s4fgg" event={"ID":"754c550f-6079-4f6d-8f18-af4ebddfdf61","Type":"ContainerDied","Data":"aec92e10d4cb243e09ecb9e45935ad305c576ca83a2923a10cd6ad13d78de63b"} Mar 09 19:56:04 crc kubenswrapper[4750]: I0309 19:56:04.870921 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1"} Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.190332 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.259784 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551436-s4fgg" Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.409348 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxnpc\" (UniqueName: \"kubernetes.io/projected/754c550f-6079-4f6d-8f18-af4ebddfdf61-kube-api-access-wxnpc\") pod \"754c550f-6079-4f6d-8f18-af4ebddfdf61\" (UID: \"754c550f-6079-4f6d-8f18-af4ebddfdf61\") " Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.420866 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/754c550f-6079-4f6d-8f18-af4ebddfdf61-kube-api-access-wxnpc" (OuterVolumeSpecName: "kube-api-access-wxnpc") pod "754c550f-6079-4f6d-8f18-af4ebddfdf61" (UID: "754c550f-6079-4f6d-8f18-af4ebddfdf61"). InnerVolumeSpecName "kube-api-access-wxnpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.512360 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxnpc\" (UniqueName: \"kubernetes.io/projected/754c550f-6079-4f6d-8f18-af4ebddfdf61-kube-api-access-wxnpc\") on node \"crc\" DevicePath \"\"" Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.879493 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551436-s4fgg" event={"ID":"754c550f-6079-4f6d-8f18-af4ebddfdf61","Type":"ContainerDied","Data":"36a59350874f2afae09ffd739490661c9cf13be0297bbe714b4669db2340acd2"} Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.879808 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36a59350874f2afae09ffd739490661c9cf13be0297bbe714b4669db2340acd2" Mar 09 19:56:05 crc kubenswrapper[4750]: I0309 19:56:05.879513 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551436-s4fgg" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.352528 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551430-wnfl2"] Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.366306 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551430-wnfl2"] Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.621728 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zdt9v"] Mar 09 19:56:06 crc kubenswrapper[4750]: E0309 19:56:06.622603 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="754c550f-6079-4f6d-8f18-af4ebddfdf61" containerName="oc" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.622679 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="754c550f-6079-4f6d-8f18-af4ebddfdf61" containerName="oc" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.623240 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="754c550f-6079-4f6d-8f18-af4ebddfdf61" containerName="oc" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.626914 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.633921 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdt9v"] Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.738463 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-catalog-content\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.738658 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpbxh\" (UniqueName: \"kubernetes.io/projected/87280ed7-1246-4c12-9cf1-0966b41cf447-kube-api-access-zpbxh\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.738772 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-utilities\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.840499 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-utilities\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.840555 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-catalog-content\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.840676 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpbxh\" (UniqueName: \"kubernetes.io/projected/87280ed7-1246-4c12-9cf1-0966b41cf447-kube-api-access-zpbxh\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.841081 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-catalog-content\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.841260 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-utilities\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.860962 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpbxh\" (UniqueName: \"kubernetes.io/projected/87280ed7-1246-4c12-9cf1-0966b41cf447-kube-api-access-zpbxh\") pod \"redhat-marketplace-zdt9v\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.895761 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" exitCode=1 Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.895858 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1"} Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.895930 4750 scope.go:117] "RemoveContainer" containerID="03993c06a0a06ff054cc803b04b9032667aa319a8b1f78afc7b01b360f25e28d" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.897779 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:56:06 crc kubenswrapper[4750]: E0309 19:56:06.898389 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:56:06 crc kubenswrapper[4750]: I0309 19:56:06.959330 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:07 crc kubenswrapper[4750]: I0309 19:56:07.383321 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b150d127-1ec5-4ec6-8ff3-1541b355c39d" path="/var/lib/kubelet/pods/b150d127-1ec5-4ec6-8ff3-1541b355c39d/volumes" Mar 09 19:56:07 crc kubenswrapper[4750]: I0309 19:56:07.454131 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdt9v"] Mar 09 19:56:07 crc kubenswrapper[4750]: I0309 19:56:07.920293 4750 generic.go:334] "Generic (PLEG): container finished" podID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerID="98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680" exitCode=0 Mar 09 19:56:07 crc kubenswrapper[4750]: I0309 19:56:07.920593 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdt9v" event={"ID":"87280ed7-1246-4c12-9cf1-0966b41cf447","Type":"ContainerDied","Data":"98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680"} Mar 09 19:56:07 crc kubenswrapper[4750]: I0309 19:56:07.920615 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdt9v" event={"ID":"87280ed7-1246-4c12-9cf1-0966b41cf447","Type":"ContainerStarted","Data":"d9e747ec6bf55ba1f08ee64c30405bdee0ffb28addf38f66f9f95bb80d9e7a04"} Mar 09 19:56:08 crc kubenswrapper[4750]: I0309 19:56:08.189339 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:56:08 crc kubenswrapper[4750]: I0309 19:56:08.189944 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:56:08 crc kubenswrapper[4750]: E0309 19:56:08.190271 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:56:09 crc kubenswrapper[4750]: I0309 19:56:09.945742 4750 generic.go:334] "Generic (PLEG): container finished" podID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerID="fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd" exitCode=0 Mar 09 19:56:09 crc kubenswrapper[4750]: I0309 19:56:09.946056 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdt9v" event={"ID":"87280ed7-1246-4c12-9cf1-0966b41cf447","Type":"ContainerDied","Data":"fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd"} Mar 09 19:56:10 crc kubenswrapper[4750]: I0309 19:56:10.189709 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 19:56:10 crc kubenswrapper[4750]: I0309 19:56:10.190851 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:56:10 crc kubenswrapper[4750]: E0309 19:56:10.197959 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:56:10 crc kubenswrapper[4750]: I0309 19:56:10.958336 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdt9v" event={"ID":"87280ed7-1246-4c12-9cf1-0966b41cf447","Type":"ContainerStarted","Data":"521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284"} Mar 09 19:56:10 crc kubenswrapper[4750]: I0309 19:56:10.984704 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zdt9v" podStartSLOduration=2.402797906 podStartE2EDuration="4.984684262s" podCreationTimestamp="2026-03-09 19:56:06 +0000 UTC" firstStartedPulling="2026-03-09 19:56:07.922647175 +0000 UTC m=+5449.265119583" lastFinishedPulling="2026-03-09 19:56:10.504533501 +0000 UTC m=+5451.847005939" observedRunningTime="2026-03-09 19:56:10.978315679 +0000 UTC m=+5452.320788107" watchObservedRunningTime="2026-03-09 19:56:10.984684262 +0000 UTC m=+5452.327156670" Mar 09 19:56:15 crc kubenswrapper[4750]: I0309 19:56:15.374840 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:56:15 crc kubenswrapper[4750]: E0309 19:56:15.376034 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:56:16 crc kubenswrapper[4750]: I0309 19:56:16.960023 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:16 crc kubenswrapper[4750]: I0309 19:56:16.960240 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:17 crc kubenswrapper[4750]: I0309 19:56:17.029164 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:17 crc kubenswrapper[4750]: I0309 19:56:17.125202 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:17 crc kubenswrapper[4750]: I0309 19:56:17.274789 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdt9v"] Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.065415 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zdt9v" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="registry-server" containerID="cri-o://521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284" gracePeriod=2 Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.673069 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.768373 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-catalog-content\") pod \"87280ed7-1246-4c12-9cf1-0966b41cf447\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.768503 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-utilities\") pod \"87280ed7-1246-4c12-9cf1-0966b41cf447\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.768709 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpbxh\" (UniqueName: \"kubernetes.io/projected/87280ed7-1246-4c12-9cf1-0966b41cf447-kube-api-access-zpbxh\") pod \"87280ed7-1246-4c12-9cf1-0966b41cf447\" (UID: \"87280ed7-1246-4c12-9cf1-0966b41cf447\") " Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.769507 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-utilities" (OuterVolumeSpecName: "utilities") pod "87280ed7-1246-4c12-9cf1-0966b41cf447" (UID: "87280ed7-1246-4c12-9cf1-0966b41cf447"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.791058 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87280ed7-1246-4c12-9cf1-0966b41cf447-kube-api-access-zpbxh" (OuterVolumeSpecName: "kube-api-access-zpbxh") pod "87280ed7-1246-4c12-9cf1-0966b41cf447" (UID: "87280ed7-1246-4c12-9cf1-0966b41cf447"). InnerVolumeSpecName "kube-api-access-zpbxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.808329 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87280ed7-1246-4c12-9cf1-0966b41cf447" (UID: "87280ed7-1246-4c12-9cf1-0966b41cf447"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.871436 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpbxh\" (UniqueName: \"kubernetes.io/projected/87280ed7-1246-4c12-9cf1-0966b41cf447-kube-api-access-zpbxh\") on node \"crc\" DevicePath \"\"" Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.871462 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 19:56:19 crc kubenswrapper[4750]: I0309 19:56:19.871472 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87280ed7-1246-4c12-9cf1-0966b41cf447-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.076249 4750 generic.go:334] "Generic (PLEG): container finished" podID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerID="521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284" exitCode=0 Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.076306 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdt9v" event={"ID":"87280ed7-1246-4c12-9cf1-0966b41cf447","Type":"ContainerDied","Data":"521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284"} Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.076318 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdt9v" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.076337 4750 scope.go:117] "RemoveContainer" containerID="521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.076328 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdt9v" event={"ID":"87280ed7-1246-4c12-9cf1-0966b41cf447","Type":"ContainerDied","Data":"d9e747ec6bf55ba1f08ee64c30405bdee0ffb28addf38f66f9f95bb80d9e7a04"} Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.111064 4750 scope.go:117] "RemoveContainer" containerID="fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.139871 4750 scope.go:117] "RemoveContainer" containerID="98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.171911 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdt9v"] Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.182957 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdt9v"] Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.225561 4750 scope.go:117] "RemoveContainer" containerID="521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284" Mar 09 19:56:20 crc kubenswrapper[4750]: E0309 19:56:20.226476 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284\": container with ID starting with 521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284 not found: ID does not exist" containerID="521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.226530 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284"} err="failed to get container status \"521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284\": rpc error: code = NotFound desc = could not find container \"521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284\": container with ID starting with 521f96e2fa2ab89808c205cd8c427c62689f3b23329d004185b59a74864a2284 not found: ID does not exist" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.226578 4750 scope.go:117] "RemoveContainer" containerID="fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd" Mar 09 19:56:20 crc kubenswrapper[4750]: E0309 19:56:20.226857 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd\": container with ID starting with fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd not found: ID does not exist" containerID="fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.226879 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd"} err="failed to get container status \"fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd\": rpc error: code = NotFound desc = could not find container \"fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd\": container with ID starting with fc020d68824521e6e599963c0926e6704b0bffefd1772fc7fbda3e16444598fd not found: ID does not exist" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.226898 4750 scope.go:117] "RemoveContainer" containerID="98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680" Mar 09 19:56:20 crc kubenswrapper[4750]: E0309 19:56:20.227140 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680\": container with ID starting with 98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680 not found: ID does not exist" containerID="98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680" Mar 09 19:56:20 crc kubenswrapper[4750]: I0309 19:56:20.227179 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680"} err="failed to get container status \"98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680\": rpc error: code = NotFound desc = could not find container \"98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680\": container with ID starting with 98a74cd146f036fecf6e10d29f4d67c2c7cc1486d8a195fecf365af65e7a0680 not found: ID does not exist" Mar 09 19:56:21 crc kubenswrapper[4750]: I0309 19:56:21.389319 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" path="/var/lib/kubelet/pods/87280ed7-1246-4c12-9cf1-0966b41cf447/volumes" Mar 09 19:56:25 crc kubenswrapper[4750]: I0309 19:56:25.375353 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:56:25 crc kubenswrapper[4750]: E0309 19:56:25.379340 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:56:29 crc kubenswrapper[4750]: I0309 19:56:29.379557 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:56:29 crc kubenswrapper[4750]: E0309 19:56:29.380223 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:56:39 crc kubenswrapper[4750]: I0309 19:56:39.380884 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:56:39 crc kubenswrapper[4750]: E0309 19:56:39.386242 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:56:40 crc kubenswrapper[4750]: I0309 19:56:40.373977 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:56:40 crc kubenswrapper[4750]: E0309 19:56:40.374747 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:56:50 crc kubenswrapper[4750]: I0309 19:56:50.137672 4750 scope.go:117] "RemoveContainer" containerID="0ebcaccb501c5f98438b7b26063abee356b333cfe18196f1f25962a62ef03a36" Mar 09 19:56:52 crc kubenswrapper[4750]: I0309 19:56:52.374604 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:56:52 crc kubenswrapper[4750]: I0309 19:56:52.375400 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:56:52 crc kubenswrapper[4750]: E0309 19:56:52.375514 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:56:52 crc kubenswrapper[4750]: E0309 19:56:52.375936 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:57:05 crc kubenswrapper[4750]: I0309 19:57:05.374811 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:57:05 crc kubenswrapper[4750]: I0309 19:57:05.375696 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:57:05 crc kubenswrapper[4750]: E0309 19:57:05.376076 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:57:05 crc kubenswrapper[4750]: E0309 19:57:05.376280 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:57:16 crc kubenswrapper[4750]: I0309 19:57:16.373337 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:57:16 crc kubenswrapper[4750]: I0309 19:57:16.373984 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:57:16 crc kubenswrapper[4750]: E0309 19:57:16.374315 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:57:16 crc kubenswrapper[4750]: E0309 19:57:16.374501 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:57:29 crc kubenswrapper[4750]: I0309 19:57:29.388786 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:57:29 crc kubenswrapper[4750]: E0309 19:57:29.390179 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:57:31 crc kubenswrapper[4750]: I0309 19:57:31.375031 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:57:31 crc kubenswrapper[4750]: E0309 19:57:31.375862 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:57:40 crc kubenswrapper[4750]: I0309 19:57:40.373774 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:57:40 crc kubenswrapper[4750]: E0309 19:57:40.374877 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:57:43 crc kubenswrapper[4750]: I0309 19:57:43.373707 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:57:43 crc kubenswrapper[4750]: E0309 19:57:43.374705 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:57:54 crc kubenswrapper[4750]: I0309 19:57:54.374772 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:57:54 crc kubenswrapper[4750]: I0309 19:57:54.375350 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:57:54 crc kubenswrapper[4750]: E0309 19:57:54.375556 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:57:54 crc kubenswrapper[4750]: E0309 19:57:54.375664 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.182834 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551438-2hsxg"] Mar 09 19:58:00 crc kubenswrapper[4750]: E0309 19:58:00.183859 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="registry-server" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.183872 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="registry-server" Mar 09 19:58:00 crc kubenswrapper[4750]: E0309 19:58:00.183883 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="extract-utilities" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.183889 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="extract-utilities" Mar 09 19:58:00 crc kubenswrapper[4750]: E0309 19:58:00.183895 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="extract-content" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.183901 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="extract-content" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.184104 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="87280ed7-1246-4c12-9cf1-0966b41cf447" containerName="registry-server" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.186232 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551438-2hsxg" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.188226 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.189183 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.189386 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.202101 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551438-2hsxg"] Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.319035 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxtk9\" (UniqueName: \"kubernetes.io/projected/b695ff27-b930-4af1-ae3c-e0a5fc1a9841-kube-api-access-cxtk9\") pod \"auto-csr-approver-29551438-2hsxg\" (UID: \"b695ff27-b930-4af1-ae3c-e0a5fc1a9841\") " pod="openshift-infra/auto-csr-approver-29551438-2hsxg" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.421909 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxtk9\" (UniqueName: \"kubernetes.io/projected/b695ff27-b930-4af1-ae3c-e0a5fc1a9841-kube-api-access-cxtk9\") pod \"auto-csr-approver-29551438-2hsxg\" (UID: \"b695ff27-b930-4af1-ae3c-e0a5fc1a9841\") " pod="openshift-infra/auto-csr-approver-29551438-2hsxg" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.455213 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxtk9\" (UniqueName: \"kubernetes.io/projected/b695ff27-b930-4af1-ae3c-e0a5fc1a9841-kube-api-access-cxtk9\") pod \"auto-csr-approver-29551438-2hsxg\" (UID: \"b695ff27-b930-4af1-ae3c-e0a5fc1a9841\") " pod="openshift-infra/auto-csr-approver-29551438-2hsxg" Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.508855 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551438-2hsxg" Mar 09 19:58:00 crc kubenswrapper[4750]: W0309 19:58:00.843820 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb695ff27_b930_4af1_ae3c_e0a5fc1a9841.slice/crio-b6ae337b5aaca5a4c8ef7d0581f4d76fcdfda8957ae6e55bdbea9518678b3490 WatchSource:0}: Error finding container b6ae337b5aaca5a4c8ef7d0581f4d76fcdfda8957ae6e55bdbea9518678b3490: Status 404 returned error can't find the container with id b6ae337b5aaca5a4c8ef7d0581f4d76fcdfda8957ae6e55bdbea9518678b3490 Mar 09 19:58:00 crc kubenswrapper[4750]: I0309 19:58:00.844213 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551438-2hsxg"] Mar 09 19:58:01 crc kubenswrapper[4750]: I0309 19:58:01.343489 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551438-2hsxg" event={"ID":"b695ff27-b930-4af1-ae3c-e0a5fc1a9841","Type":"ContainerStarted","Data":"b6ae337b5aaca5a4c8ef7d0581f4d76fcdfda8957ae6e55bdbea9518678b3490"} Mar 09 19:58:03 crc kubenswrapper[4750]: I0309 19:58:03.365798 4750 generic.go:334] "Generic (PLEG): container finished" podID="b695ff27-b930-4af1-ae3c-e0a5fc1a9841" containerID="2d7df9f62ecbb32e29fc1e1112822221547c55dde48e9205ba2af2140588bc48" exitCode=0 Mar 09 19:58:03 crc kubenswrapper[4750]: I0309 19:58:03.365872 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551438-2hsxg" event={"ID":"b695ff27-b930-4af1-ae3c-e0a5fc1a9841","Type":"ContainerDied","Data":"2d7df9f62ecbb32e29fc1e1112822221547c55dde48e9205ba2af2140588bc48"} Mar 09 19:58:04 crc kubenswrapper[4750]: I0309 19:58:04.731557 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551438-2hsxg" Mar 09 19:58:04 crc kubenswrapper[4750]: I0309 19:58:04.834976 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxtk9\" (UniqueName: \"kubernetes.io/projected/b695ff27-b930-4af1-ae3c-e0a5fc1a9841-kube-api-access-cxtk9\") pod \"b695ff27-b930-4af1-ae3c-e0a5fc1a9841\" (UID: \"b695ff27-b930-4af1-ae3c-e0a5fc1a9841\") " Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.374953 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:58:05 crc kubenswrapper[4750]: E0309 19:58:05.375669 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.390908 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551438-2hsxg" Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.395571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551438-2hsxg" event={"ID":"b695ff27-b930-4af1-ae3c-e0a5fc1a9841","Type":"ContainerDied","Data":"b6ae337b5aaca5a4c8ef7d0581f4d76fcdfda8957ae6e55bdbea9518678b3490"} Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.395622 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6ae337b5aaca5a4c8ef7d0581f4d76fcdfda8957ae6e55bdbea9518678b3490" Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.462299 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b695ff27-b930-4af1-ae3c-e0a5fc1a9841-kube-api-access-cxtk9" (OuterVolumeSpecName: "kube-api-access-cxtk9") pod "b695ff27-b930-4af1-ae3c-e0a5fc1a9841" (UID: "b695ff27-b930-4af1-ae3c-e0a5fc1a9841"). InnerVolumeSpecName "kube-api-access-cxtk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.553006 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxtk9\" (UniqueName: \"kubernetes.io/projected/b695ff27-b930-4af1-ae3c-e0a5fc1a9841-kube-api-access-cxtk9\") on node \"crc\" DevicePath \"\"" Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.846010 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551432-2mjmk"] Mar 09 19:58:05 crc kubenswrapper[4750]: I0309 19:58:05.877439 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551432-2mjmk"] Mar 09 19:58:06 crc kubenswrapper[4750]: I0309 19:58:06.375189 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:58:06 crc kubenswrapper[4750]: E0309 19:58:06.375901 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:58:07 crc kubenswrapper[4750]: I0309 19:58:07.391662 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5262a0c-eea0-42c8-9b41-20d2045e593d" path="/var/lib/kubelet/pods/f5262a0c-eea0-42c8-9b41-20d2045e593d/volumes" Mar 09 19:58:16 crc kubenswrapper[4750]: I0309 19:58:16.374537 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:58:16 crc kubenswrapper[4750]: E0309 19:58:16.375567 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:58:19 crc kubenswrapper[4750]: I0309 19:58:19.392691 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:58:19 crc kubenswrapper[4750]: E0309 19:58:19.394684 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:58:27 crc kubenswrapper[4750]: I0309 19:58:27.373519 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:58:27 crc kubenswrapper[4750]: E0309 19:58:27.374321 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:58:32 crc kubenswrapper[4750]: I0309 19:58:32.373565 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:58:32 crc kubenswrapper[4750]: E0309 19:58:32.374920 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:58:41 crc kubenswrapper[4750]: I0309 19:58:41.374146 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:58:41 crc kubenswrapper[4750]: E0309 19:58:41.375413 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:58:47 crc kubenswrapper[4750]: I0309 19:58:47.373999 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:58:47 crc kubenswrapper[4750]: E0309 19:58:47.375207 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:58:50 crc kubenswrapper[4750]: I0309 19:58:50.314671 4750 scope.go:117] "RemoveContainer" containerID="78e18030fadfc0ef37e45f340a312c160da6c8e354be867e652389be5bfc4fe1" Mar 09 19:58:53 crc kubenswrapper[4750]: I0309 19:58:53.375369 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:58:53 crc kubenswrapper[4750]: E0309 19:58:53.376105 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:58:58 crc kubenswrapper[4750]: I0309 19:58:58.373820 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:58:58 crc kubenswrapper[4750]: E0309 19:58:58.374875 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:59:04 crc kubenswrapper[4750]: I0309 19:59:04.373500 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:59:04 crc kubenswrapper[4750]: E0309 19:59:04.374552 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:59:13 crc kubenswrapper[4750]: I0309 19:59:13.422087 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:59:13 crc kubenswrapper[4750]: E0309 19:59:13.423684 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 19:59:19 crc kubenswrapper[4750]: I0309 19:59:19.388212 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:59:19 crc kubenswrapper[4750]: E0309 19:59:19.388944 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:59:27 crc kubenswrapper[4750]: I0309 19:59:27.374366 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 19:59:28 crc kubenswrapper[4750]: I0309 19:59:28.449251 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"10b7bd4c3a60a8e0251271952503f2a5d9934f7b6d607ba8a0e9ba88efd62371"} Mar 09 19:59:30 crc kubenswrapper[4750]: I0309 19:59:30.373927 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:59:30 crc kubenswrapper[4750]: E0309 19:59:30.374803 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:59:42 crc kubenswrapper[4750]: I0309 19:59:42.374601 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:59:42 crc kubenswrapper[4750]: E0309 19:59:42.375604 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 19:59:55 crc kubenswrapper[4750]: I0309 19:59:55.376369 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 19:59:55 crc kubenswrapper[4750]: E0309 19:59:55.377163 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.152104 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551440-8t64l"] Mar 09 20:00:00 crc kubenswrapper[4750]: E0309 20:00:00.155527 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b695ff27-b930-4af1-ae3c-e0a5fc1a9841" containerName="oc" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.155671 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b695ff27-b930-4af1-ae3c-e0a5fc1a9841" containerName="oc" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.156072 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b695ff27-b930-4af1-ae3c-e0a5fc1a9841" containerName="oc" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.157059 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551440-8t64l" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.159693 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.160254 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.160515 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.172242 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx"] Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.175391 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.178103 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.178357 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.189965 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551440-8t64l"] Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.209107 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x2sh\" (UniqueName: \"kubernetes.io/projected/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-kube-api-access-6x2sh\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.209207 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj54s\" (UniqueName: \"kubernetes.io/projected/ccdfffd9-b346-443a-823d-6f1fb249ce33-kube-api-access-cj54s\") pod \"auto-csr-approver-29551440-8t64l\" (UID: \"ccdfffd9-b346-443a-823d-6f1fb249ce33\") " pod="openshift-infra/auto-csr-approver-29551440-8t64l" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.209381 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-secret-volume\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.209454 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-config-volume\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.218676 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx"] Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.311362 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-secret-volume\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.311440 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-config-volume\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.311472 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x2sh\" (UniqueName: \"kubernetes.io/projected/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-kube-api-access-6x2sh\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.311513 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj54s\" (UniqueName: \"kubernetes.io/projected/ccdfffd9-b346-443a-823d-6f1fb249ce33-kube-api-access-cj54s\") pod \"auto-csr-approver-29551440-8t64l\" (UID: \"ccdfffd9-b346-443a-823d-6f1fb249ce33\") " pod="openshift-infra/auto-csr-approver-29551440-8t64l" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.312471 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-config-volume\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.324258 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-secret-volume\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.327750 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj54s\" (UniqueName: \"kubernetes.io/projected/ccdfffd9-b346-443a-823d-6f1fb249ce33-kube-api-access-cj54s\") pod \"auto-csr-approver-29551440-8t64l\" (UID: \"ccdfffd9-b346-443a-823d-6f1fb249ce33\") " pod="openshift-infra/auto-csr-approver-29551440-8t64l" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.335259 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x2sh\" (UniqueName: \"kubernetes.io/projected/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-kube-api-access-6x2sh\") pod \"collect-profiles-29551440-vltkx\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.494544 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551440-8t64l" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.511709 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:00 crc kubenswrapper[4750]: I0309 20:00:00.813430 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551440-8t64l"] Mar 09 20:00:00 crc kubenswrapper[4750]: W0309 20:00:00.815996 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccdfffd9_b346_443a_823d_6f1fb249ce33.slice/crio-584eb75fa30a531a084dde68cd960966e7b9e33c3d449ef6405a9c2318f450e2 WatchSource:0}: Error finding container 584eb75fa30a531a084dde68cd960966e7b9e33c3d449ef6405a9c2318f450e2: Status 404 returned error can't find the container with id 584eb75fa30a531a084dde68cd960966e7b9e33c3d449ef6405a9c2318f450e2 Mar 09 20:00:01 crc kubenswrapper[4750]: I0309 20:00:01.100703 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx"] Mar 09 20:00:01 crc kubenswrapper[4750]: I0309 20:00:01.822436 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551440-8t64l" event={"ID":"ccdfffd9-b346-443a-823d-6f1fb249ce33","Type":"ContainerStarted","Data":"584eb75fa30a531a084dde68cd960966e7b9e33c3d449ef6405a9c2318f450e2"} Mar 09 20:00:01 crc kubenswrapper[4750]: I0309 20:00:01.824365 4750 generic.go:334] "Generic (PLEG): container finished" podID="32c9b7e4-8216-450e-8cd1-02b68cd48ec7" containerID="878e22d3e7094fdbd96715063ff61c3afdff01ed4293c8ad3640f03589051f04" exitCode=0 Mar 09 20:00:01 crc kubenswrapper[4750]: I0309 20:00:01.824431 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" event={"ID":"32c9b7e4-8216-450e-8cd1-02b68cd48ec7","Type":"ContainerDied","Data":"878e22d3e7094fdbd96715063ff61c3afdff01ed4293c8ad3640f03589051f04"} Mar 09 20:00:01 crc kubenswrapper[4750]: I0309 20:00:01.824479 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" event={"ID":"32c9b7e4-8216-450e-8cd1-02b68cd48ec7","Type":"ContainerStarted","Data":"9424868dd1451fa16c6a422e15e341379aadf401d2eb058b03246d05c7aef25c"} Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.227718 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.283836 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-config-volume\") pod \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.284019 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-secret-volume\") pod \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.284180 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x2sh\" (UniqueName: \"kubernetes.io/projected/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-kube-api-access-6x2sh\") pod \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\" (UID: \"32c9b7e4-8216-450e-8cd1-02b68cd48ec7\") " Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.287138 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-config-volume" (OuterVolumeSpecName: "config-volume") pod "32c9b7e4-8216-450e-8cd1-02b68cd48ec7" (UID: "32c9b7e4-8216-450e-8cd1-02b68cd48ec7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.291468 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-kube-api-access-6x2sh" (OuterVolumeSpecName: "kube-api-access-6x2sh") pod "32c9b7e4-8216-450e-8cd1-02b68cd48ec7" (UID: "32c9b7e4-8216-450e-8cd1-02b68cd48ec7"). InnerVolumeSpecName "kube-api-access-6x2sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.291478 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "32c9b7e4-8216-450e-8cd1-02b68cd48ec7" (UID: "32c9b7e4-8216-450e-8cd1-02b68cd48ec7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.388552 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.388931 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x2sh\" (UniqueName: \"kubernetes.io/projected/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-kube-api-access-6x2sh\") on node \"crc\" DevicePath \"\"" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.388946 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c9b7e4-8216-450e-8cd1-02b68cd48ec7-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.861133 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" event={"ID":"32c9b7e4-8216-450e-8cd1-02b68cd48ec7","Type":"ContainerDied","Data":"9424868dd1451fa16c6a422e15e341379aadf401d2eb058b03246d05c7aef25c"} Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.861210 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9424868dd1451fa16c6a422e15e341379aadf401d2eb058b03246d05c7aef25c" Mar 09 20:00:03 crc kubenswrapper[4750]: I0309 20:00:03.861246 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551440-vltkx" Mar 09 20:00:04 crc kubenswrapper[4750]: I0309 20:00:04.344442 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f"] Mar 09 20:00:04 crc kubenswrapper[4750]: I0309 20:00:04.359260 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551395-2fp5f"] Mar 09 20:00:04 crc kubenswrapper[4750]: I0309 20:00:04.871527 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551440-8t64l" event={"ID":"ccdfffd9-b346-443a-823d-6f1fb249ce33","Type":"ContainerStarted","Data":"d0b2134f2194e3442440ab585b863ae55327e6978dd0701cc78bfc13c241efb8"} Mar 09 20:00:04 crc kubenswrapper[4750]: I0309 20:00:04.886795 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551440-8t64l" podStartSLOduration=1.3212952470000001 podStartE2EDuration="4.886775313s" podCreationTimestamp="2026-03-09 20:00:00 +0000 UTC" firstStartedPulling="2026-03-09 20:00:00.817558759 +0000 UTC m=+5682.160031157" lastFinishedPulling="2026-03-09 20:00:04.383038805 +0000 UTC m=+5685.725511223" observedRunningTime="2026-03-09 20:00:04.886517665 +0000 UTC m=+5686.228990073" watchObservedRunningTime="2026-03-09 20:00:04.886775313 +0000 UTC m=+5686.229247711" Mar 09 20:00:05 crc kubenswrapper[4750]: I0309 20:00:05.387921 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d42a7252-53c5-411e-8b76-20926501ab47" path="/var/lib/kubelet/pods/d42a7252-53c5-411e-8b76-20926501ab47/volumes" Mar 09 20:00:05 crc kubenswrapper[4750]: I0309 20:00:05.887290 4750 generic.go:334] "Generic (PLEG): container finished" podID="ccdfffd9-b346-443a-823d-6f1fb249ce33" containerID="d0b2134f2194e3442440ab585b863ae55327e6978dd0701cc78bfc13c241efb8" exitCode=0 Mar 09 20:00:05 crc kubenswrapper[4750]: I0309 20:00:05.887369 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551440-8t64l" event={"ID":"ccdfffd9-b346-443a-823d-6f1fb249ce33","Type":"ContainerDied","Data":"d0b2134f2194e3442440ab585b863ae55327e6978dd0701cc78bfc13c241efb8"} Mar 09 20:00:06 crc kubenswrapper[4750]: I0309 20:00:06.373616 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 20:00:06 crc kubenswrapper[4750]: E0309 20:00:06.374243 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.311223 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551440-8t64l" Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.392730 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj54s\" (UniqueName: \"kubernetes.io/projected/ccdfffd9-b346-443a-823d-6f1fb249ce33-kube-api-access-cj54s\") pod \"ccdfffd9-b346-443a-823d-6f1fb249ce33\" (UID: \"ccdfffd9-b346-443a-823d-6f1fb249ce33\") " Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.405845 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccdfffd9-b346-443a-823d-6f1fb249ce33-kube-api-access-cj54s" (OuterVolumeSpecName: "kube-api-access-cj54s") pod "ccdfffd9-b346-443a-823d-6f1fb249ce33" (UID: "ccdfffd9-b346-443a-823d-6f1fb249ce33"). InnerVolumeSpecName "kube-api-access-cj54s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.495525 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj54s\" (UniqueName: \"kubernetes.io/projected/ccdfffd9-b346-443a-823d-6f1fb249ce33-kube-api-access-cj54s\") on node \"crc\" DevicePath \"\"" Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.913046 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551440-8t64l" event={"ID":"ccdfffd9-b346-443a-823d-6f1fb249ce33","Type":"ContainerDied","Data":"584eb75fa30a531a084dde68cd960966e7b9e33c3d449ef6405a9c2318f450e2"} Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.913397 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="584eb75fa30a531a084dde68cd960966e7b9e33c3d449ef6405a9c2318f450e2" Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.913287 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551440-8t64l" Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.959922 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551434-2gxz9"] Mar 09 20:00:07 crc kubenswrapper[4750]: I0309 20:00:07.966866 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551434-2gxz9"] Mar 09 20:00:09 crc kubenswrapper[4750]: I0309 20:00:09.392267 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e14dabf4-c85c-481b-8025-d4874c8b1a8a" path="/var/lib/kubelet/pods/e14dabf4-c85c-481b-8025-d4874c8b1a8a/volumes" Mar 09 20:00:17 crc kubenswrapper[4750]: I0309 20:00:17.373833 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 20:00:17 crc kubenswrapper[4750]: E0309 20:00:17.376042 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:00:30 crc kubenswrapper[4750]: I0309 20:00:30.375552 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 20:00:30 crc kubenswrapper[4750]: E0309 20:00:30.376594 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:00:44 crc kubenswrapper[4750]: I0309 20:00:44.372948 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 20:00:44 crc kubenswrapper[4750]: E0309 20:00:44.373568 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.040252 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rg77b"] Mar 09 20:00:50 crc kubenswrapper[4750]: E0309 20:00:50.041496 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c9b7e4-8216-450e-8cd1-02b68cd48ec7" containerName="collect-profiles" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.041538 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c9b7e4-8216-450e-8cd1-02b68cd48ec7" containerName="collect-profiles" Mar 09 20:00:50 crc kubenswrapper[4750]: E0309 20:00:50.041574 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccdfffd9-b346-443a-823d-6f1fb249ce33" containerName="oc" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.041586 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccdfffd9-b346-443a-823d-6f1fb249ce33" containerName="oc" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.041931 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccdfffd9-b346-443a-823d-6f1fb249ce33" containerName="oc" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.041961 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c9b7e4-8216-450e-8cd1-02b68cd48ec7" containerName="collect-profiles" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.044270 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.055049 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rg77b"] Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.174144 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-utilities\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.174514 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltv6x\" (UniqueName: \"kubernetes.io/projected/abf88d14-4878-441c-9a3e-2528c20db090-kube-api-access-ltv6x\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.174543 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-catalog-content\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.276193 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltv6x\" (UniqueName: \"kubernetes.io/projected/abf88d14-4878-441c-9a3e-2528c20db090-kube-api-access-ltv6x\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.276264 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-catalog-content\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.276427 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-utilities\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.276900 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-catalog-content\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.277135 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-utilities\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.306946 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltv6x\" (UniqueName: \"kubernetes.io/projected/abf88d14-4878-441c-9a3e-2528c20db090-kube-api-access-ltv6x\") pod \"community-operators-rg77b\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.369991 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.964652 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rg77b"] Mar 09 20:00:50 crc kubenswrapper[4750]: I0309 20:00:50.990094 4750 scope.go:117] "RemoveContainer" containerID="7838a59444f0d8d58c3612f05b8fc5e01b920b063da0f726328ec6bc624aa173" Mar 09 20:00:51 crc kubenswrapper[4750]: I0309 20:00:51.298428 4750 scope.go:117] "RemoveContainer" containerID="bbbb6ec93f7b95cfa24c9cfd98b584ddb549fcd9be22f020b40929d30b2cdca0" Mar 09 20:00:51 crc kubenswrapper[4750]: I0309 20:00:51.406339 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg77b" event={"ID":"abf88d14-4878-441c-9a3e-2528c20db090","Type":"ContainerStarted","Data":"520654686ef53cc7361900ec087880670d8072772e4970e34c010d4596712bf6"} Mar 09 20:00:51 crc kubenswrapper[4750]: I0309 20:00:51.572366 4750 scope.go:117] "RemoveContainer" containerID="9803ec38e2477ee94c5282a9e4eb36a747f9ff901fd7b68a533a53c7e9f1f7b5" Mar 09 20:00:51 crc kubenswrapper[4750]: I0309 20:00:51.601154 4750 scope.go:117] "RemoveContainer" containerID="3680ca273c2c4465f9cae1e9c7187c9edfdaa651a28f13050068f97698f864ca" Mar 09 20:00:51 crc kubenswrapper[4750]: I0309 20:00:51.686335 4750 scope.go:117] "RemoveContainer" containerID="b24d417c6f76a416f23654c1ddf46b1d0b72a69749ee6bad769126c64e1a69c6" Mar 09 20:00:52 crc kubenswrapper[4750]: I0309 20:00:52.414570 4750 generic.go:334] "Generic (PLEG): container finished" podID="abf88d14-4878-441c-9a3e-2528c20db090" containerID="7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe" exitCode=0 Mar 09 20:00:52 crc kubenswrapper[4750]: I0309 20:00:52.414664 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg77b" event={"ID":"abf88d14-4878-441c-9a3e-2528c20db090","Type":"ContainerDied","Data":"7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe"} Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.239663 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sj4px"] Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.242297 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.257352 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sj4px"] Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.345881 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwmgn\" (UniqueName: \"kubernetes.io/projected/12af313c-21ad-4617-86c7-15a0bfa7177c-kube-api-access-qwmgn\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.346177 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-utilities\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.346382 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-catalog-content\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.426159 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg77b" event={"ID":"abf88d14-4878-441c-9a3e-2528c20db090","Type":"ContainerStarted","Data":"58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868"} Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.448734 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwmgn\" (UniqueName: \"kubernetes.io/projected/12af313c-21ad-4617-86c7-15a0bfa7177c-kube-api-access-qwmgn\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.449218 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-utilities\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.449675 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-catalog-content\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.449896 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-utilities\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.450076 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-catalog-content\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.483770 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwmgn\" (UniqueName: \"kubernetes.io/projected/12af313c-21ad-4617-86c7-15a0bfa7177c-kube-api-access-qwmgn\") pod \"certified-operators-sj4px\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:53 crc kubenswrapper[4750]: I0309 20:00:53.571674 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:00:54 crc kubenswrapper[4750]: I0309 20:00:54.124895 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sj4px"] Mar 09 20:00:54 crc kubenswrapper[4750]: I0309 20:00:54.444766 4750 generic.go:334] "Generic (PLEG): container finished" podID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerID="3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d" exitCode=0 Mar 09 20:00:54 crc kubenswrapper[4750]: I0309 20:00:54.447613 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj4px" event={"ID":"12af313c-21ad-4617-86c7-15a0bfa7177c","Type":"ContainerDied","Data":"3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d"} Mar 09 20:00:54 crc kubenswrapper[4750]: I0309 20:00:54.447659 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj4px" event={"ID":"12af313c-21ad-4617-86c7-15a0bfa7177c","Type":"ContainerStarted","Data":"c7b6978416af6521b3b43b8307d9f39fc0bc43edfc3beff01836268a9eb7e514"} Mar 09 20:00:55 crc kubenswrapper[4750]: I0309 20:00:55.376417 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 20:00:55 crc kubenswrapper[4750]: E0309 20:00:55.376884 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:00:55 crc kubenswrapper[4750]: I0309 20:00:55.453773 4750 generic.go:334] "Generic (PLEG): container finished" podID="abf88d14-4878-441c-9a3e-2528c20db090" containerID="58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868" exitCode=0 Mar 09 20:00:55 crc kubenswrapper[4750]: I0309 20:00:55.453811 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg77b" event={"ID":"abf88d14-4878-441c-9a3e-2528c20db090","Type":"ContainerDied","Data":"58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868"} Mar 09 20:00:56 crc kubenswrapper[4750]: I0309 20:00:56.471842 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg77b" event={"ID":"abf88d14-4878-441c-9a3e-2528c20db090","Type":"ContainerStarted","Data":"6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da"} Mar 09 20:00:56 crc kubenswrapper[4750]: I0309 20:00:56.474650 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj4px" event={"ID":"12af313c-21ad-4617-86c7-15a0bfa7177c","Type":"ContainerStarted","Data":"2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1"} Mar 09 20:00:56 crc kubenswrapper[4750]: I0309 20:00:56.495943 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rg77b" podStartSLOduration=2.850157881 podStartE2EDuration="6.495922455s" podCreationTimestamp="2026-03-09 20:00:50 +0000 UTC" firstStartedPulling="2026-03-09 20:00:52.416202858 +0000 UTC m=+5733.758675266" lastFinishedPulling="2026-03-09 20:00:56.061967442 +0000 UTC m=+5737.404439840" observedRunningTime="2026-03-09 20:00:56.490088517 +0000 UTC m=+5737.832560945" watchObservedRunningTime="2026-03-09 20:00:56.495922455 +0000 UTC m=+5737.838394863" Mar 09 20:00:58 crc kubenswrapper[4750]: I0309 20:00:58.493297 4750 generic.go:334] "Generic (PLEG): container finished" podID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerID="2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1" exitCode=0 Mar 09 20:00:58 crc kubenswrapper[4750]: I0309 20:00:58.493386 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj4px" event={"ID":"12af313c-21ad-4617-86c7-15a0bfa7177c","Type":"ContainerDied","Data":"2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1"} Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.210529 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29551441-8sgzm"] Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.212471 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.232094 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29551441-8sgzm"] Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.303768 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-fernet-keys\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.303811 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-combined-ca-bundle\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.303883 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-config-data\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.303973 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szxpl\" (UniqueName: \"kubernetes.io/projected/5137a335-d7e5-4aff-bb01-19eb8d5600d6-kube-api-access-szxpl\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.370603 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.370679 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.406173 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szxpl\" (UniqueName: \"kubernetes.io/projected/5137a335-d7e5-4aff-bb01-19eb8d5600d6-kube-api-access-szxpl\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.406284 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-fernet-keys\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.406312 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-combined-ca-bundle\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.406373 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-config-data\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.415866 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-config-data\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.419512 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-combined-ca-bundle\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.434486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szxpl\" (UniqueName: \"kubernetes.io/projected/5137a335-d7e5-4aff-bb01-19eb8d5600d6-kube-api-access-szxpl\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.438354 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-fernet-keys\") pod \"keystone-cron-29551441-8sgzm\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.465805 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.515677 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj4px" event={"ID":"12af313c-21ad-4617-86c7-15a0bfa7177c","Type":"ContainerStarted","Data":"6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2"} Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.537926 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:00 crc kubenswrapper[4750]: I0309 20:01:00.550130 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sj4px" podStartSLOduration=3.078476108 podStartE2EDuration="7.550112183s" podCreationTimestamp="2026-03-09 20:00:53 +0000 UTC" firstStartedPulling="2026-03-09 20:00:54.448297572 +0000 UTC m=+5735.790769970" lastFinishedPulling="2026-03-09 20:00:58.919933617 +0000 UTC m=+5740.262406045" observedRunningTime="2026-03-09 20:01:00.541971463 +0000 UTC m=+5741.884443871" watchObservedRunningTime="2026-03-09 20:01:00.550112183 +0000 UTC m=+5741.892584581" Mar 09 20:01:01 crc kubenswrapper[4750]: I0309 20:01:01.040349 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29551441-8sgzm"] Mar 09 20:01:01 crc kubenswrapper[4750]: I0309 20:01:01.526047 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551441-8sgzm" event={"ID":"5137a335-d7e5-4aff-bb01-19eb8d5600d6","Type":"ContainerStarted","Data":"63a531e87628dfd855df67de2683a1e0ae730db9f39809054f03ded79e9a335f"} Mar 09 20:01:01 crc kubenswrapper[4750]: I0309 20:01:01.526348 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551441-8sgzm" event={"ID":"5137a335-d7e5-4aff-bb01-19eb8d5600d6","Type":"ContainerStarted","Data":"542cc2447fc5040702b0f5e0b9f3c2571e89419d970a50b73b6f163410883056"} Mar 09 20:01:01 crc kubenswrapper[4750]: I0309 20:01:01.545382 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29551441-8sgzm" podStartSLOduration=1.545366088 podStartE2EDuration="1.545366088s" podCreationTimestamp="2026-03-09 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 20:01:01.540467236 +0000 UTC m=+5742.882939644" watchObservedRunningTime="2026-03-09 20:01:01.545366088 +0000 UTC m=+5742.887838496" Mar 09 20:01:03 crc kubenswrapper[4750]: I0309 20:01:03.572522 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:01:03 crc kubenswrapper[4750]: I0309 20:01:03.573725 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:01:03 crc kubenswrapper[4750]: I0309 20:01:03.645126 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:01:05 crc kubenswrapper[4750]: I0309 20:01:05.570880 4750 generic.go:334] "Generic (PLEG): container finished" podID="5137a335-d7e5-4aff-bb01-19eb8d5600d6" containerID="63a531e87628dfd855df67de2683a1e0ae730db9f39809054f03ded79e9a335f" exitCode=0 Mar 09 20:01:05 crc kubenswrapper[4750]: I0309 20:01:05.570911 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551441-8sgzm" event={"ID":"5137a335-d7e5-4aff-bb01-19eb8d5600d6","Type":"ContainerDied","Data":"63a531e87628dfd855df67de2683a1e0ae730db9f39809054f03ded79e9a335f"} Mar 09 20:01:06 crc kubenswrapper[4750]: I0309 20:01:06.964419 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.069014 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-combined-ca-bundle\") pod \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.069067 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-config-data\") pod \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.069143 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szxpl\" (UniqueName: \"kubernetes.io/projected/5137a335-d7e5-4aff-bb01-19eb8d5600d6-kube-api-access-szxpl\") pod \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.069177 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-fernet-keys\") pod \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\" (UID: \"5137a335-d7e5-4aff-bb01-19eb8d5600d6\") " Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.077482 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5137a335-d7e5-4aff-bb01-19eb8d5600d6" (UID: "5137a335-d7e5-4aff-bb01-19eb8d5600d6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.081016 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5137a335-d7e5-4aff-bb01-19eb8d5600d6-kube-api-access-szxpl" (OuterVolumeSpecName: "kube-api-access-szxpl") pod "5137a335-d7e5-4aff-bb01-19eb8d5600d6" (UID: "5137a335-d7e5-4aff-bb01-19eb8d5600d6"). InnerVolumeSpecName "kube-api-access-szxpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.135990 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5137a335-d7e5-4aff-bb01-19eb8d5600d6" (UID: "5137a335-d7e5-4aff-bb01-19eb8d5600d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.136366 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-config-data" (OuterVolumeSpecName: "config-data") pod "5137a335-d7e5-4aff-bb01-19eb8d5600d6" (UID: "5137a335-d7e5-4aff-bb01-19eb8d5600d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.171611 4750 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.171666 4750 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.171680 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szxpl\" (UniqueName: \"kubernetes.io/projected/5137a335-d7e5-4aff-bb01-19eb8d5600d6-kube-api-access-szxpl\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.171692 4750 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5137a335-d7e5-4aff-bb01-19eb8d5600d6-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.597092 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29551441-8sgzm" event={"ID":"5137a335-d7e5-4aff-bb01-19eb8d5600d6","Type":"ContainerDied","Data":"542cc2447fc5040702b0f5e0b9f3c2571e89419d970a50b73b6f163410883056"} Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.597383 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="542cc2447fc5040702b0f5e0b9f3c2571e89419d970a50b73b6f163410883056" Mar 09 20:01:07 crc kubenswrapper[4750]: I0309 20:01:07.597194 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29551441-8sgzm" Mar 09 20:01:10 crc kubenswrapper[4750]: I0309 20:01:10.374945 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 20:01:10 crc kubenswrapper[4750]: I0309 20:01:10.465391 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:01:11 crc kubenswrapper[4750]: I0309 20:01:11.634760 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6"} Mar 09 20:01:13 crc kubenswrapper[4750]: I0309 20:01:13.646510 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:01:13 crc kubenswrapper[4750]: I0309 20:01:13.657615 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" exitCode=1 Mar 09 20:01:13 crc kubenswrapper[4750]: I0309 20:01:13.657680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6"} Mar 09 20:01:13 crc kubenswrapper[4750]: I0309 20:01:13.657717 4750 scope.go:117] "RemoveContainer" containerID="85bd7ef7032f1c2599db2caa252e021d6b398bc9ba5cedcd9d770b54315661c1" Mar 09 20:01:13 crc kubenswrapper[4750]: I0309 20:01:13.658452 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:01:13 crc kubenswrapper[4750]: E0309 20:01:13.658845 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:01:14 crc kubenswrapper[4750]: I0309 20:01:14.189754 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:01:14 crc kubenswrapper[4750]: I0309 20:01:14.687273 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:01:14 crc kubenswrapper[4750]: E0309 20:01:14.687783 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:01:15 crc kubenswrapper[4750]: I0309 20:01:15.189843 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:01:15 crc kubenswrapper[4750]: I0309 20:01:15.190194 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:01:15 crc kubenswrapper[4750]: I0309 20:01:15.696540 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:01:15 crc kubenswrapper[4750]: E0309 20:01:15.696983 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.027941 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sj4px"] Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.028296 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sj4px" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="registry-server" containerID="cri-o://6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2" gracePeriod=2 Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.546643 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.668581 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-catalog-content\") pod \"12af313c-21ad-4617-86c7-15a0bfa7177c\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.668882 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwmgn\" (UniqueName: \"kubernetes.io/projected/12af313c-21ad-4617-86c7-15a0bfa7177c-kube-api-access-qwmgn\") pod \"12af313c-21ad-4617-86c7-15a0bfa7177c\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.668992 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-utilities\") pod \"12af313c-21ad-4617-86c7-15a0bfa7177c\" (UID: \"12af313c-21ad-4617-86c7-15a0bfa7177c\") " Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.670329 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-utilities" (OuterVolumeSpecName: "utilities") pod "12af313c-21ad-4617-86c7-15a0bfa7177c" (UID: "12af313c-21ad-4617-86c7-15a0bfa7177c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.690869 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12af313c-21ad-4617-86c7-15a0bfa7177c-kube-api-access-qwmgn" (OuterVolumeSpecName: "kube-api-access-qwmgn") pod "12af313c-21ad-4617-86c7-15a0bfa7177c" (UID: "12af313c-21ad-4617-86c7-15a0bfa7177c"). InnerVolumeSpecName "kube-api-access-qwmgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.749365 4750 generic.go:334] "Generic (PLEG): container finished" podID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerID="6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2" exitCode=0 Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.749411 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj4px" event={"ID":"12af313c-21ad-4617-86c7-15a0bfa7177c","Type":"ContainerDied","Data":"6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2"} Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.749439 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj4px" event={"ID":"12af313c-21ad-4617-86c7-15a0bfa7177c","Type":"ContainerDied","Data":"c7b6978416af6521b3b43b8307d9f39fc0bc43edfc3beff01836268a9eb7e514"} Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.749454 4750 scope.go:117] "RemoveContainer" containerID="6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.749614 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj4px" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.762432 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12af313c-21ad-4617-86c7-15a0bfa7177c" (UID: "12af313c-21ad-4617-86c7-15a0bfa7177c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.770975 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.771013 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12af313c-21ad-4617-86c7-15a0bfa7177c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.771026 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwmgn\" (UniqueName: \"kubernetes.io/projected/12af313c-21ad-4617-86c7-15a0bfa7177c-kube-api-access-qwmgn\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.783730 4750 scope.go:117] "RemoveContainer" containerID="2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.823884 4750 scope.go:117] "RemoveContainer" containerID="3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.866624 4750 scope.go:117] "RemoveContainer" containerID="6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2" Mar 09 20:01:18 crc kubenswrapper[4750]: E0309 20:01:18.867089 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2\": container with ID starting with 6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2 not found: ID does not exist" containerID="6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.867120 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2"} err="failed to get container status \"6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2\": rpc error: code = NotFound desc = could not find container \"6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2\": container with ID starting with 6bb0632be3b9abf878238f24ac37e2f300c1a6cb6f51cdca5503a754fb1017d2 not found: ID does not exist" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.867143 4750 scope.go:117] "RemoveContainer" containerID="2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1" Mar 09 20:01:18 crc kubenswrapper[4750]: E0309 20:01:18.867397 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1\": container with ID starting with 2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1 not found: ID does not exist" containerID="2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.867456 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1"} err="failed to get container status \"2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1\": rpc error: code = NotFound desc = could not find container \"2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1\": container with ID starting with 2379471948c78158131cb3f4dd20c971610c0ad87e5d5659053c4e6595e271c1 not found: ID does not exist" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.867494 4750 scope.go:117] "RemoveContainer" containerID="3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d" Mar 09 20:01:18 crc kubenswrapper[4750]: E0309 20:01:18.867981 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d\": container with ID starting with 3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d not found: ID does not exist" containerID="3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d" Mar 09 20:01:18 crc kubenswrapper[4750]: I0309 20:01:18.868005 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d"} err="failed to get container status \"3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d\": rpc error: code = NotFound desc = could not find container \"3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d\": container with ID starting with 3fbc660ba9b4c95e3910a16a016ce151a614b0db037d10b358901dfc93faae3d not found: ID does not exist" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.086368 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sj4px"] Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.099912 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sj4px"] Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.223259 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rg77b"] Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.223962 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rg77b" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="registry-server" containerID="cri-o://6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da" gracePeriod=2 Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.393872 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" path="/var/lib/kubelet/pods/12af313c-21ad-4617-86c7-15a0bfa7177c/volumes" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.715683 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.784290 4750 generic.go:334] "Generic (PLEG): container finished" podID="abf88d14-4878-441c-9a3e-2528c20db090" containerID="6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da" exitCode=0 Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.784388 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg77b" event={"ID":"abf88d14-4878-441c-9a3e-2528c20db090","Type":"ContainerDied","Data":"6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da"} Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.784426 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg77b" event={"ID":"abf88d14-4878-441c-9a3e-2528c20db090","Type":"ContainerDied","Data":"520654686ef53cc7361900ec087880670d8072772e4970e34c010d4596712bf6"} Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.784460 4750 scope.go:117] "RemoveContainer" containerID="6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.784706 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg77b" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.824448 4750 scope.go:117] "RemoveContainer" containerID="58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.852917 4750 scope.go:117] "RemoveContainer" containerID="7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.899768 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-catalog-content\") pod \"abf88d14-4878-441c-9a3e-2528c20db090\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.899898 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltv6x\" (UniqueName: \"kubernetes.io/projected/abf88d14-4878-441c-9a3e-2528c20db090-kube-api-access-ltv6x\") pod \"abf88d14-4878-441c-9a3e-2528c20db090\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.900809 4750 scope.go:117] "RemoveContainer" containerID="6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.900905 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-utilities\") pod \"abf88d14-4878-441c-9a3e-2528c20db090\" (UID: \"abf88d14-4878-441c-9a3e-2528c20db090\") " Mar 09 20:01:19 crc kubenswrapper[4750]: E0309 20:01:19.901265 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da\": container with ID starting with 6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da not found: ID does not exist" containerID="6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.901293 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da"} err="failed to get container status \"6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da\": rpc error: code = NotFound desc = could not find container \"6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da\": container with ID starting with 6706ac751bbc31edb4bac4f8b2120688aed00102658f6289e9e77e57f02ac0da not found: ID does not exist" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.901311 4750 scope.go:117] "RemoveContainer" containerID="58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.901829 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-utilities" (OuterVolumeSpecName: "utilities") pod "abf88d14-4878-441c-9a3e-2528c20db090" (UID: "abf88d14-4878-441c-9a3e-2528c20db090"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:01:19 crc kubenswrapper[4750]: E0309 20:01:19.901837 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868\": container with ID starting with 58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868 not found: ID does not exist" containerID="58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.901985 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868"} err="failed to get container status \"58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868\": rpc error: code = NotFound desc = could not find container \"58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868\": container with ID starting with 58e1bbb5e5a0fb83f9b897f47597bf5578829382614ff4530f6a6886772f7868 not found: ID does not exist" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.902058 4750 scope.go:117] "RemoveContainer" containerID="7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe" Mar 09 20:01:19 crc kubenswrapper[4750]: E0309 20:01:19.903036 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe\": container with ID starting with 7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe not found: ID does not exist" containerID="7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.903118 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe"} err="failed to get container status \"7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe\": rpc error: code = NotFound desc = could not find container \"7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe\": container with ID starting with 7b3e3d92b2ecd0cee7e257ce535865ceec518245151e94c6646c14f06a4a40fe not found: ID does not exist" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.912185 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf88d14-4878-441c-9a3e-2528c20db090-kube-api-access-ltv6x" (OuterVolumeSpecName: "kube-api-access-ltv6x") pod "abf88d14-4878-441c-9a3e-2528c20db090" (UID: "abf88d14-4878-441c-9a3e-2528c20db090"). InnerVolumeSpecName "kube-api-access-ltv6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:01:19 crc kubenswrapper[4750]: I0309 20:01:19.949000 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abf88d14-4878-441c-9a3e-2528c20db090" (UID: "abf88d14-4878-441c-9a3e-2528c20db090"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:01:20 crc kubenswrapper[4750]: I0309 20:01:20.003919 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:20 crc kubenswrapper[4750]: I0309 20:01:20.004160 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf88d14-4878-441c-9a3e-2528c20db090-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:20 crc kubenswrapper[4750]: I0309 20:01:20.004254 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltv6x\" (UniqueName: \"kubernetes.io/projected/abf88d14-4878-441c-9a3e-2528c20db090-kube-api-access-ltv6x\") on node \"crc\" DevicePath \"\"" Mar 09 20:01:20 crc kubenswrapper[4750]: I0309 20:01:20.145129 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rg77b"] Mar 09 20:01:20 crc kubenswrapper[4750]: I0309 20:01:20.159260 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rg77b"] Mar 09 20:01:21 crc kubenswrapper[4750]: I0309 20:01:21.394592 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf88d14-4878-441c-9a3e-2528c20db090" path="/var/lib/kubelet/pods/abf88d14-4878-441c-9a3e-2528c20db090/volumes" Mar 09 20:01:27 crc kubenswrapper[4750]: I0309 20:01:27.374133 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:01:27 crc kubenswrapper[4750]: E0309 20:01:27.375299 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:01:38 crc kubenswrapper[4750]: I0309 20:01:38.374955 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:01:38 crc kubenswrapper[4750]: E0309 20:01:38.378203 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:01:49 crc kubenswrapper[4750]: I0309 20:01:49.380306 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:01:49 crc kubenswrapper[4750]: E0309 20:01:49.381240 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:01:51 crc kubenswrapper[4750]: I0309 20:01:51.743147 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:01:51 crc kubenswrapper[4750]: I0309 20:01:51.743525 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.172482 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551442-ldtvn"] Mar 09 20:02:00 crc kubenswrapper[4750]: E0309 20:02:00.173913 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="registry-server" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.173934 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="registry-server" Mar 09 20:02:00 crc kubenswrapper[4750]: E0309 20:02:00.173956 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="extract-content" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.173968 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="extract-content" Mar 09 20:02:00 crc kubenswrapper[4750]: E0309 20:02:00.173997 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="extract-content" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174007 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="extract-content" Mar 09 20:02:00 crc kubenswrapper[4750]: E0309 20:02:00.174026 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="extract-utilities" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174037 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="extract-utilities" Mar 09 20:02:00 crc kubenswrapper[4750]: E0309 20:02:00.174064 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="registry-server" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174075 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="registry-server" Mar 09 20:02:00 crc kubenswrapper[4750]: E0309 20:02:00.174104 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5137a335-d7e5-4aff-bb01-19eb8d5600d6" containerName="keystone-cron" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174113 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5137a335-d7e5-4aff-bb01-19eb8d5600d6" containerName="keystone-cron" Mar 09 20:02:00 crc kubenswrapper[4750]: E0309 20:02:00.174142 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="extract-utilities" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174152 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="extract-utilities" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174462 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="12af313c-21ad-4617-86c7-15a0bfa7177c" containerName="registry-server" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174490 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5137a335-d7e5-4aff-bb01-19eb8d5600d6" containerName="keystone-cron" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.174517 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf88d14-4878-441c-9a3e-2528c20db090" containerName="registry-server" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.175619 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551442-ldtvn" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.178822 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.179219 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.179366 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.195409 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551442-ldtvn"] Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.221858 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcp4b\" (UniqueName: \"kubernetes.io/projected/5c8f040d-1d00-42c3-94e5-bec6babbbe48-kube-api-access-rcp4b\") pod \"auto-csr-approver-29551442-ldtvn\" (UID: \"5c8f040d-1d00-42c3-94e5-bec6babbbe48\") " pod="openshift-infra/auto-csr-approver-29551442-ldtvn" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.323773 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcp4b\" (UniqueName: \"kubernetes.io/projected/5c8f040d-1d00-42c3-94e5-bec6babbbe48-kube-api-access-rcp4b\") pod \"auto-csr-approver-29551442-ldtvn\" (UID: \"5c8f040d-1d00-42c3-94e5-bec6babbbe48\") " pod="openshift-infra/auto-csr-approver-29551442-ldtvn" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.354573 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcp4b\" (UniqueName: \"kubernetes.io/projected/5c8f040d-1d00-42c3-94e5-bec6babbbe48-kube-api-access-rcp4b\") pod \"auto-csr-approver-29551442-ldtvn\" (UID: \"5c8f040d-1d00-42c3-94e5-bec6babbbe48\") " pod="openshift-infra/auto-csr-approver-29551442-ldtvn" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.515517 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551442-ldtvn" Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.859608 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551442-ldtvn"] Mar 09 20:02:00 crc kubenswrapper[4750]: I0309 20:02:00.877452 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 20:02:01 crc kubenswrapper[4750]: I0309 20:02:01.324808 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551442-ldtvn" event={"ID":"5c8f040d-1d00-42c3-94e5-bec6babbbe48","Type":"ContainerStarted","Data":"fd8ee7b8dfa4f19c31f39f1c4152c957faa6bd1367817e0a0898ad60b8b883e1"} Mar 09 20:02:01 crc kubenswrapper[4750]: I0309 20:02:01.374336 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:02:01 crc kubenswrapper[4750]: E0309 20:02:01.374719 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:02:02 crc kubenswrapper[4750]: I0309 20:02:02.336206 4750 generic.go:334] "Generic (PLEG): container finished" podID="5c8f040d-1d00-42c3-94e5-bec6babbbe48" containerID="a361870b2a77fa92467d58b17a999624ac65935640121eec7535ff3f005d72dc" exitCode=0 Mar 09 20:02:02 crc kubenswrapper[4750]: I0309 20:02:02.336287 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551442-ldtvn" event={"ID":"5c8f040d-1d00-42c3-94e5-bec6babbbe48","Type":"ContainerDied","Data":"a361870b2a77fa92467d58b17a999624ac65935640121eec7535ff3f005d72dc"} Mar 09 20:02:03 crc kubenswrapper[4750]: I0309 20:02:03.781947 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551442-ldtvn" Mar 09 20:02:03 crc kubenswrapper[4750]: I0309 20:02:03.909713 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcp4b\" (UniqueName: \"kubernetes.io/projected/5c8f040d-1d00-42c3-94e5-bec6babbbe48-kube-api-access-rcp4b\") pod \"5c8f040d-1d00-42c3-94e5-bec6babbbe48\" (UID: \"5c8f040d-1d00-42c3-94e5-bec6babbbe48\") " Mar 09 20:02:03 crc kubenswrapper[4750]: I0309 20:02:03.916241 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c8f040d-1d00-42c3-94e5-bec6babbbe48-kube-api-access-rcp4b" (OuterVolumeSpecName: "kube-api-access-rcp4b") pod "5c8f040d-1d00-42c3-94e5-bec6babbbe48" (UID: "5c8f040d-1d00-42c3-94e5-bec6babbbe48"). InnerVolumeSpecName "kube-api-access-rcp4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:02:04 crc kubenswrapper[4750]: I0309 20:02:04.012648 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcp4b\" (UniqueName: \"kubernetes.io/projected/5c8f040d-1d00-42c3-94e5-bec6babbbe48-kube-api-access-rcp4b\") on node \"crc\" DevicePath \"\"" Mar 09 20:02:04 crc kubenswrapper[4750]: I0309 20:02:04.384288 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551442-ldtvn" Mar 09 20:02:04 crc kubenswrapper[4750]: I0309 20:02:04.384198 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551442-ldtvn" event={"ID":"5c8f040d-1d00-42c3-94e5-bec6babbbe48","Type":"ContainerDied","Data":"fd8ee7b8dfa4f19c31f39f1c4152c957faa6bd1367817e0a0898ad60b8b883e1"} Mar 09 20:02:04 crc kubenswrapper[4750]: I0309 20:02:04.392039 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd8ee7b8dfa4f19c31f39f1c4152c957faa6bd1367817e0a0898ad60b8b883e1" Mar 09 20:02:04 crc kubenswrapper[4750]: I0309 20:02:04.891076 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551436-s4fgg"] Mar 09 20:02:04 crc kubenswrapper[4750]: I0309 20:02:04.905448 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551436-s4fgg"] Mar 09 20:02:05 crc kubenswrapper[4750]: I0309 20:02:05.399682 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="754c550f-6079-4f6d-8f18-af4ebddfdf61" path="/var/lib/kubelet/pods/754c550f-6079-4f6d-8f18-af4ebddfdf61/volumes" Mar 09 20:02:14 crc kubenswrapper[4750]: I0309 20:02:14.373617 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:02:14 crc kubenswrapper[4750]: E0309 20:02:14.374407 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:02:21 crc kubenswrapper[4750]: I0309 20:02:21.744219 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:02:21 crc kubenswrapper[4750]: I0309 20:02:21.745211 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:02:29 crc kubenswrapper[4750]: I0309 20:02:29.403898 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:02:29 crc kubenswrapper[4750]: E0309 20:02:29.404842 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:02:40 crc kubenswrapper[4750]: I0309 20:02:40.373877 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:02:40 crc kubenswrapper[4750]: E0309 20:02:40.374950 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.743900 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.744498 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.744550 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.745465 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"10b7bd4c3a60a8e0251271952503f2a5d9934f7b6d607ba8a0e9ba88efd62371"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.745529 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://10b7bd4c3a60a8e0251271952503f2a5d9934f7b6d607ba8a0e9ba88efd62371" gracePeriod=600 Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.851504 4750 scope.go:117] "RemoveContainer" containerID="aec92e10d4cb243e09ecb9e45935ad305c576ca83a2923a10cd6ad13d78de63b" Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.947689 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="10b7bd4c3a60a8e0251271952503f2a5d9934f7b6d607ba8a0e9ba88efd62371" exitCode=0 Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.947740 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"10b7bd4c3a60a8e0251271952503f2a5d9934f7b6d607ba8a0e9ba88efd62371"} Mar 09 20:02:51 crc kubenswrapper[4750]: I0309 20:02:51.947818 4750 scope.go:117] "RemoveContainer" containerID="c658b075f2906cf6de7db533288d1926651db43d52f317106bf0bc2ad5dcb768" Mar 09 20:02:52 crc kubenswrapper[4750]: I0309 20:02:52.965026 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d"} Mar 09 20:02:53 crc kubenswrapper[4750]: I0309 20:02:53.374658 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:02:53 crc kubenswrapper[4750]: E0309 20:02:53.375161 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:03:08 crc kubenswrapper[4750]: I0309 20:03:08.373350 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:03:08 crc kubenswrapper[4750]: E0309 20:03:08.374723 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:03:20 crc kubenswrapper[4750]: I0309 20:03:20.373448 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:03:20 crc kubenswrapper[4750]: E0309 20:03:20.375273 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:03:33 crc kubenswrapper[4750]: I0309 20:03:33.374690 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:03:33 crc kubenswrapper[4750]: E0309 20:03:33.376061 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:03:45 crc kubenswrapper[4750]: I0309 20:03:45.373886 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:03:45 crc kubenswrapper[4750]: E0309 20:03:45.375744 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:03:56 crc kubenswrapper[4750]: I0309 20:03:56.375314 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:03:56 crc kubenswrapper[4750]: E0309 20:03:56.377980 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.165836 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551444-stpgq"] Mar 09 20:04:00 crc kubenswrapper[4750]: E0309 20:04:00.166900 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8f040d-1d00-42c3-94e5-bec6babbbe48" containerName="oc" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.166918 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8f040d-1d00-42c3-94e5-bec6babbbe48" containerName="oc" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.167198 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8f040d-1d00-42c3-94e5-bec6babbbe48" containerName="oc" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.168046 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551444-stpgq" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.174208 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.174268 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.174443 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.223329 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551444-stpgq"] Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.335664 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbl9p\" (UniqueName: \"kubernetes.io/projected/3e704c07-6189-469e-87fd-10f711cbc618-kube-api-access-qbl9p\") pod \"auto-csr-approver-29551444-stpgq\" (UID: \"3e704c07-6189-469e-87fd-10f711cbc618\") " pod="openshift-infra/auto-csr-approver-29551444-stpgq" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.437605 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbl9p\" (UniqueName: \"kubernetes.io/projected/3e704c07-6189-469e-87fd-10f711cbc618-kube-api-access-qbl9p\") pod \"auto-csr-approver-29551444-stpgq\" (UID: \"3e704c07-6189-469e-87fd-10f711cbc618\") " pod="openshift-infra/auto-csr-approver-29551444-stpgq" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.465079 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbl9p\" (UniqueName: \"kubernetes.io/projected/3e704c07-6189-469e-87fd-10f711cbc618-kube-api-access-qbl9p\") pod \"auto-csr-approver-29551444-stpgq\" (UID: \"3e704c07-6189-469e-87fd-10f711cbc618\") " pod="openshift-infra/auto-csr-approver-29551444-stpgq" Mar 09 20:04:00 crc kubenswrapper[4750]: I0309 20:04:00.519447 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551444-stpgq" Mar 09 20:04:01 crc kubenswrapper[4750]: I0309 20:04:01.030612 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551444-stpgq"] Mar 09 20:04:01 crc kubenswrapper[4750]: W0309 20:04:01.041305 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e704c07_6189_469e_87fd_10f711cbc618.slice/crio-cf403ec16f633981d9ad4c9b7cf4e0b1220012d331dc2cf6a1fae7fae65f19a8 WatchSource:0}: Error finding container cf403ec16f633981d9ad4c9b7cf4e0b1220012d331dc2cf6a1fae7fae65f19a8: Status 404 returned error can't find the container with id cf403ec16f633981d9ad4c9b7cf4e0b1220012d331dc2cf6a1fae7fae65f19a8 Mar 09 20:04:01 crc kubenswrapper[4750]: I0309 20:04:01.771653 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551444-stpgq" event={"ID":"3e704c07-6189-469e-87fd-10f711cbc618","Type":"ContainerStarted","Data":"cf403ec16f633981d9ad4c9b7cf4e0b1220012d331dc2cf6a1fae7fae65f19a8"} Mar 09 20:04:02 crc kubenswrapper[4750]: I0309 20:04:02.786436 4750 generic.go:334] "Generic (PLEG): container finished" podID="3e704c07-6189-469e-87fd-10f711cbc618" containerID="0e91a60a9bdded50c0d7315cfe08b58aa8ac24cb8adb4b758cb4b553d18a4626" exitCode=0 Mar 09 20:04:02 crc kubenswrapper[4750]: I0309 20:04:02.786567 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551444-stpgq" event={"ID":"3e704c07-6189-469e-87fd-10f711cbc618","Type":"ContainerDied","Data":"0e91a60a9bdded50c0d7315cfe08b58aa8ac24cb8adb4b758cb4b553d18a4626"} Mar 09 20:04:04 crc kubenswrapper[4750]: I0309 20:04:04.208704 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551444-stpgq" Mar 09 20:04:04 crc kubenswrapper[4750]: I0309 20:04:04.323142 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbl9p\" (UniqueName: \"kubernetes.io/projected/3e704c07-6189-469e-87fd-10f711cbc618-kube-api-access-qbl9p\") pod \"3e704c07-6189-469e-87fd-10f711cbc618\" (UID: \"3e704c07-6189-469e-87fd-10f711cbc618\") " Mar 09 20:04:04 crc kubenswrapper[4750]: I0309 20:04:04.329001 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e704c07-6189-469e-87fd-10f711cbc618-kube-api-access-qbl9p" (OuterVolumeSpecName: "kube-api-access-qbl9p") pod "3e704c07-6189-469e-87fd-10f711cbc618" (UID: "3e704c07-6189-469e-87fd-10f711cbc618"). InnerVolumeSpecName "kube-api-access-qbl9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:04:04 crc kubenswrapper[4750]: I0309 20:04:04.426890 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbl9p\" (UniqueName: \"kubernetes.io/projected/3e704c07-6189-469e-87fd-10f711cbc618-kube-api-access-qbl9p\") on node \"crc\" DevicePath \"\"" Mar 09 20:04:04 crc kubenswrapper[4750]: I0309 20:04:04.810703 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551444-stpgq" event={"ID":"3e704c07-6189-469e-87fd-10f711cbc618","Type":"ContainerDied","Data":"cf403ec16f633981d9ad4c9b7cf4e0b1220012d331dc2cf6a1fae7fae65f19a8"} Mar 09 20:04:04 crc kubenswrapper[4750]: I0309 20:04:04.810743 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf403ec16f633981d9ad4c9b7cf4e0b1220012d331dc2cf6a1fae7fae65f19a8" Mar 09 20:04:04 crc kubenswrapper[4750]: I0309 20:04:04.810804 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551444-stpgq" Mar 09 20:04:05 crc kubenswrapper[4750]: I0309 20:04:05.277463 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551438-2hsxg"] Mar 09 20:04:05 crc kubenswrapper[4750]: I0309 20:04:05.286000 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551438-2hsxg"] Mar 09 20:04:05 crc kubenswrapper[4750]: I0309 20:04:05.403795 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b695ff27-b930-4af1-ae3c-e0a5fc1a9841" path="/var/lib/kubelet/pods/b695ff27-b930-4af1-ae3c-e0a5fc1a9841/volumes" Mar 09 20:04:11 crc kubenswrapper[4750]: I0309 20:04:11.374061 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:04:11 crc kubenswrapper[4750]: E0309 20:04:11.374974 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:04:25 crc kubenswrapper[4750]: I0309 20:04:25.378402 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:04:25 crc kubenswrapper[4750]: E0309 20:04:25.380151 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:04:40 crc kubenswrapper[4750]: I0309 20:04:40.373123 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:04:40 crc kubenswrapper[4750]: E0309 20:04:40.374940 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:04:51 crc kubenswrapper[4750]: I0309 20:04:51.374830 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:04:51 crc kubenswrapper[4750]: E0309 20:04:51.376286 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:04:52 crc kubenswrapper[4750]: I0309 20:04:52.085333 4750 scope.go:117] "RemoveContainer" containerID="2d7df9f62ecbb32e29fc1e1112822221547c55dde48e9205ba2af2140588bc48" Mar 09 20:05:03 crc kubenswrapper[4750]: I0309 20:05:03.373443 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:05:03 crc kubenswrapper[4750]: E0309 20:05:03.374391 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:05:18 crc kubenswrapper[4750]: I0309 20:05:18.373394 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:05:18 crc kubenswrapper[4750]: E0309 20:05:18.374590 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:05:21 crc kubenswrapper[4750]: I0309 20:05:21.744053 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:05:21 crc kubenswrapper[4750]: I0309 20:05:21.744613 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:05:33 crc kubenswrapper[4750]: I0309 20:05:33.373311 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:05:33 crc kubenswrapper[4750]: E0309 20:05:33.375370 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:05:45 crc kubenswrapper[4750]: I0309 20:05:45.373507 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:05:45 crc kubenswrapper[4750]: E0309 20:05:45.374310 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:05:51 crc kubenswrapper[4750]: I0309 20:05:51.743355 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:05:51 crc kubenswrapper[4750]: I0309 20:05:51.743926 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:05:57 crc kubenswrapper[4750]: I0309 20:05:57.374152 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:05:57 crc kubenswrapper[4750]: E0309 20:05:57.375029 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.157030 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551446-cqtrz"] Mar 09 20:06:00 crc kubenswrapper[4750]: E0309 20:06:00.157996 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e704c07-6189-469e-87fd-10f711cbc618" containerName="oc" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.158011 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e704c07-6189-469e-87fd-10f711cbc618" containerName="oc" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.158259 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e704c07-6189-469e-87fd-10f711cbc618" containerName="oc" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.159031 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551446-cqtrz" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.161612 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.161857 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.162085 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.174241 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551446-cqtrz"] Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.313912 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4bww\" (UniqueName: \"kubernetes.io/projected/2f1ae607-f58d-4b0f-a484-b7c40889adde-kube-api-access-n4bww\") pod \"auto-csr-approver-29551446-cqtrz\" (UID: \"2f1ae607-f58d-4b0f-a484-b7c40889adde\") " pod="openshift-infra/auto-csr-approver-29551446-cqtrz" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.416834 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4bww\" (UniqueName: \"kubernetes.io/projected/2f1ae607-f58d-4b0f-a484-b7c40889adde-kube-api-access-n4bww\") pod \"auto-csr-approver-29551446-cqtrz\" (UID: \"2f1ae607-f58d-4b0f-a484-b7c40889adde\") " pod="openshift-infra/auto-csr-approver-29551446-cqtrz" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.439280 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4bww\" (UniqueName: \"kubernetes.io/projected/2f1ae607-f58d-4b0f-a484-b7c40889adde-kube-api-access-n4bww\") pod \"auto-csr-approver-29551446-cqtrz\" (UID: \"2f1ae607-f58d-4b0f-a484-b7c40889adde\") " pod="openshift-infra/auto-csr-approver-29551446-cqtrz" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.477094 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551446-cqtrz" Mar 09 20:06:00 crc kubenswrapper[4750]: I0309 20:06:00.952066 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551446-cqtrz"] Mar 09 20:06:01 crc kubenswrapper[4750]: I0309 20:06:01.064986 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551446-cqtrz" event={"ID":"2f1ae607-f58d-4b0f-a484-b7c40889adde","Type":"ContainerStarted","Data":"cae9d60569e70abdabc15154f82bdb9b666e58463ad64127403350ae0c28b9c1"} Mar 09 20:06:03 crc kubenswrapper[4750]: I0309 20:06:03.082738 4750 generic.go:334] "Generic (PLEG): container finished" podID="2f1ae607-f58d-4b0f-a484-b7c40889adde" containerID="7ea1a4f9921c57cac076c9dd63d7e0247057c909fe0a8dd09019c48f1e42c52d" exitCode=0 Mar 09 20:06:03 crc kubenswrapper[4750]: I0309 20:06:03.082800 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551446-cqtrz" event={"ID":"2f1ae607-f58d-4b0f-a484-b7c40889adde","Type":"ContainerDied","Data":"7ea1a4f9921c57cac076c9dd63d7e0247057c909fe0a8dd09019c48f1e42c52d"} Mar 09 20:06:04 crc kubenswrapper[4750]: I0309 20:06:04.457396 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551446-cqtrz" Mar 09 20:06:04 crc kubenswrapper[4750]: I0309 20:06:04.608875 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4bww\" (UniqueName: \"kubernetes.io/projected/2f1ae607-f58d-4b0f-a484-b7c40889adde-kube-api-access-n4bww\") pod \"2f1ae607-f58d-4b0f-a484-b7c40889adde\" (UID: \"2f1ae607-f58d-4b0f-a484-b7c40889adde\") " Mar 09 20:06:04 crc kubenswrapper[4750]: I0309 20:06:04.615996 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f1ae607-f58d-4b0f-a484-b7c40889adde-kube-api-access-n4bww" (OuterVolumeSpecName: "kube-api-access-n4bww") pod "2f1ae607-f58d-4b0f-a484-b7c40889adde" (UID: "2f1ae607-f58d-4b0f-a484-b7c40889adde"). InnerVolumeSpecName "kube-api-access-n4bww". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:06:04 crc kubenswrapper[4750]: I0309 20:06:04.714394 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4bww\" (UniqueName: \"kubernetes.io/projected/2f1ae607-f58d-4b0f-a484-b7c40889adde-kube-api-access-n4bww\") on node \"crc\" DevicePath \"\"" Mar 09 20:06:05 crc kubenswrapper[4750]: I0309 20:06:05.102139 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551446-cqtrz" event={"ID":"2f1ae607-f58d-4b0f-a484-b7c40889adde","Type":"ContainerDied","Data":"cae9d60569e70abdabc15154f82bdb9b666e58463ad64127403350ae0c28b9c1"} Mar 09 20:06:05 crc kubenswrapper[4750]: I0309 20:06:05.102180 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cae9d60569e70abdabc15154f82bdb9b666e58463ad64127403350ae0c28b9c1" Mar 09 20:06:05 crc kubenswrapper[4750]: I0309 20:06:05.102199 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551446-cqtrz" Mar 09 20:06:05 crc kubenswrapper[4750]: I0309 20:06:05.533301 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551440-8t64l"] Mar 09 20:06:05 crc kubenswrapper[4750]: I0309 20:06:05.542834 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551440-8t64l"] Mar 09 20:06:07 crc kubenswrapper[4750]: I0309 20:06:07.385371 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccdfffd9-b346-443a-823d-6f1fb249ce33" path="/var/lib/kubelet/pods/ccdfffd9-b346-443a-823d-6f1fb249ce33/volumes" Mar 09 20:06:10 crc kubenswrapper[4750]: I0309 20:06:10.373753 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:06:10 crc kubenswrapper[4750]: E0309 20:06:10.374741 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:06:21 crc kubenswrapper[4750]: I0309 20:06:21.744133 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:06:21 crc kubenswrapper[4750]: I0309 20:06:21.744941 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:06:21 crc kubenswrapper[4750]: I0309 20:06:21.744998 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 20:06:21 crc kubenswrapper[4750]: I0309 20:06:21.745871 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 20:06:21 crc kubenswrapper[4750]: I0309 20:06:21.745949 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" gracePeriod=600 Mar 09 20:06:21 crc kubenswrapper[4750]: E0309 20:06:21.871057 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:06:22 crc kubenswrapper[4750]: I0309 20:06:22.276575 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" exitCode=0 Mar 09 20:06:22 crc kubenswrapper[4750]: I0309 20:06:22.276677 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d"} Mar 09 20:06:22 crc kubenswrapper[4750]: I0309 20:06:22.276913 4750 scope.go:117] "RemoveContainer" containerID="10b7bd4c3a60a8e0251271952503f2a5d9934f7b6d607ba8a0e9ba88efd62371" Mar 09 20:06:22 crc kubenswrapper[4750]: I0309 20:06:22.278046 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:06:22 crc kubenswrapper[4750]: E0309 20:06:22.278350 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:06:23 crc kubenswrapper[4750]: I0309 20:06:23.373170 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:06:24 crc kubenswrapper[4750]: I0309 20:06:24.308019 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730"} Mar 09 20:06:25 crc kubenswrapper[4750]: I0309 20:06:25.189583 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:06:27 crc kubenswrapper[4750]: I0309 20:06:27.344656 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" exitCode=1 Mar 09 20:06:27 crc kubenswrapper[4750]: I0309 20:06:27.344695 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730"} Mar 09 20:06:27 crc kubenswrapper[4750]: I0309 20:06:27.344996 4750 scope.go:117] "RemoveContainer" containerID="596ab3afbe1edc9ad57dc2f4b6c6c596ead92f1afecbab0a04c03e771d81b3e6" Mar 09 20:06:27 crc kubenswrapper[4750]: I0309 20:06:27.345721 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:06:27 crc kubenswrapper[4750]: E0309 20:06:27.346136 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:06:29 crc kubenswrapper[4750]: I0309 20:06:29.189845 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:06:29 crc kubenswrapper[4750]: I0309 20:06:29.191315 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:06:29 crc kubenswrapper[4750]: E0309 20:06:29.191661 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:06:30 crc kubenswrapper[4750]: I0309 20:06:30.190295 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:06:30 crc kubenswrapper[4750]: I0309 20:06:30.191212 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:06:30 crc kubenswrapper[4750]: E0309 20:06:30.191809 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:06:33 crc kubenswrapper[4750]: I0309 20:06:33.373813 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:06:33 crc kubenswrapper[4750]: E0309 20:06:33.374332 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:06:43 crc kubenswrapper[4750]: I0309 20:06:43.374719 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:06:43 crc kubenswrapper[4750]: E0309 20:06:43.389349 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:06:47 crc kubenswrapper[4750]: I0309 20:06:47.373799 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:06:47 crc kubenswrapper[4750]: E0309 20:06:47.374675 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:06:52 crc kubenswrapper[4750]: I0309 20:06:52.227077 4750 scope.go:117] "RemoveContainer" containerID="d0b2134f2194e3442440ab585b863ae55327e6978dd0701cc78bfc13c241efb8" Mar 09 20:06:57 crc kubenswrapper[4750]: I0309 20:06:57.374999 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:06:57 crc kubenswrapper[4750]: E0309 20:06:57.376110 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:07:02 crc kubenswrapper[4750]: I0309 20:07:02.374281 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:07:02 crc kubenswrapper[4750]: E0309 20:07:02.375260 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:07:11 crc kubenswrapper[4750]: I0309 20:07:11.373478 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:07:11 crc kubenswrapper[4750]: E0309 20:07:11.375258 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:07:17 crc kubenswrapper[4750]: I0309 20:07:17.373439 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:07:17 crc kubenswrapper[4750]: E0309 20:07:17.374205 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.859384 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7xcnh"] Mar 09 20:07:22 crc kubenswrapper[4750]: E0309 20:07:22.860501 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1ae607-f58d-4b0f-a484-b7c40889adde" containerName="oc" Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.860518 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1ae607-f58d-4b0f-a484-b7c40889adde" containerName="oc" Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.860750 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1ae607-f58d-4b0f-a484-b7c40889adde" containerName="oc" Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.862803 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.882789 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xcnh"] Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.977580 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgzcp\" (UniqueName: \"kubernetes.io/projected/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-kube-api-access-fgzcp\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.977728 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-catalog-content\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:22 crc kubenswrapper[4750]: I0309 20:07:22.977769 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-utilities\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.079792 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-catalog-content\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.079842 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-utilities\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.079953 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgzcp\" (UniqueName: \"kubernetes.io/projected/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-kube-api-access-fgzcp\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.080719 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-catalog-content\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.080860 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-utilities\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.103574 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgzcp\" (UniqueName: \"kubernetes.io/projected/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-kube-api-access-fgzcp\") pod \"redhat-marketplace-7xcnh\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.191314 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.671449 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xcnh"] Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.958346 4750 generic.go:334] "Generic (PLEG): container finished" podID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerID="f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f" exitCode=0 Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.958396 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xcnh" event={"ID":"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2","Type":"ContainerDied","Data":"f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f"} Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.958424 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xcnh" event={"ID":"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2","Type":"ContainerStarted","Data":"f0747dbaa8f4c987f49f6aa3c17b5f6dbff79a2161aa83190e8baa1ec8038d75"} Mar 09 20:07:23 crc kubenswrapper[4750]: I0309 20:07:23.962041 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 20:07:25 crc kubenswrapper[4750]: I0309 20:07:25.376639 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:07:25 crc kubenswrapper[4750]: E0309 20:07:25.378506 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:07:26 crc kubenswrapper[4750]: I0309 20:07:26.073355 4750 generic.go:334] "Generic (PLEG): container finished" podID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerID="e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba" exitCode=0 Mar 09 20:07:26 crc kubenswrapper[4750]: I0309 20:07:26.073464 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xcnh" event={"ID":"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2","Type":"ContainerDied","Data":"e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba"} Mar 09 20:07:27 crc kubenswrapper[4750]: I0309 20:07:27.083772 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xcnh" event={"ID":"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2","Type":"ContainerStarted","Data":"cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d"} Mar 09 20:07:27 crc kubenswrapper[4750]: I0309 20:07:27.110193 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7xcnh" podStartSLOduration=2.505845918 podStartE2EDuration="5.110169922s" podCreationTimestamp="2026-03-09 20:07:22 +0000 UTC" firstStartedPulling="2026-03-09 20:07:23.961660152 +0000 UTC m=+6125.304132550" lastFinishedPulling="2026-03-09 20:07:26.565984146 +0000 UTC m=+6127.908456554" observedRunningTime="2026-03-09 20:07:27.098401143 +0000 UTC m=+6128.440873541" watchObservedRunningTime="2026-03-09 20:07:27.110169922 +0000 UTC m=+6128.452642330" Mar 09 20:07:31 crc kubenswrapper[4750]: I0309 20:07:31.374386 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:07:31 crc kubenswrapper[4750]: E0309 20:07:31.375262 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:07:33 crc kubenswrapper[4750]: I0309 20:07:33.191746 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:33 crc kubenswrapper[4750]: I0309 20:07:33.192049 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:33 crc kubenswrapper[4750]: I0309 20:07:33.239773 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:34 crc kubenswrapper[4750]: I0309 20:07:34.208056 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:34 crc kubenswrapper[4750]: I0309 20:07:34.273328 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xcnh"] Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.175674 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7xcnh" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="registry-server" containerID="cri-o://cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d" gracePeriod=2 Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.374293 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:07:36 crc kubenswrapper[4750]: E0309 20:07:36.374888 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.680144 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.779556 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-utilities\") pod \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.779674 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-catalog-content\") pod \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.779822 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgzcp\" (UniqueName: \"kubernetes.io/projected/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-kube-api-access-fgzcp\") pod \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\" (UID: \"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2\") " Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.780569 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-utilities" (OuterVolumeSpecName: "utilities") pod "67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" (UID: "67d2b8fe-09ce-4c0b-9e56-998bf9d070a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.785340 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-kube-api-access-fgzcp" (OuterVolumeSpecName: "kube-api-access-fgzcp") pod "67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" (UID: "67d2b8fe-09ce-4c0b-9e56-998bf9d070a2"). InnerVolumeSpecName "kube-api-access-fgzcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.806449 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" (UID: "67d2b8fe-09ce-4c0b-9e56-998bf9d070a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.882460 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.882501 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:07:36 crc kubenswrapper[4750]: I0309 20:07:36.882516 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgzcp\" (UniqueName: \"kubernetes.io/projected/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2-kube-api-access-fgzcp\") on node \"crc\" DevicePath \"\"" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.188000 4750 generic.go:334] "Generic (PLEG): container finished" podID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerID="cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d" exitCode=0 Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.188040 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xcnh" event={"ID":"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2","Type":"ContainerDied","Data":"cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d"} Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.188066 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xcnh" event={"ID":"67d2b8fe-09ce-4c0b-9e56-998bf9d070a2","Type":"ContainerDied","Data":"f0747dbaa8f4c987f49f6aa3c17b5f6dbff79a2161aa83190e8baa1ec8038d75"} Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.188083 4750 scope.go:117] "RemoveContainer" containerID="cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.188113 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xcnh" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.210085 4750 scope.go:117] "RemoveContainer" containerID="e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.250104 4750 scope.go:117] "RemoveContainer" containerID="f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.257070 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xcnh"] Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.270336 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xcnh"] Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.322837 4750 scope.go:117] "RemoveContainer" containerID="cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d" Mar 09 20:07:37 crc kubenswrapper[4750]: E0309 20:07:37.323392 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d\": container with ID starting with cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d not found: ID does not exist" containerID="cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.323479 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d"} err="failed to get container status \"cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d\": rpc error: code = NotFound desc = could not find container \"cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d\": container with ID starting with cbbee2a01f22c76c2be8e239d7416060e5019843838fd457bc65275be76b1a6d not found: ID does not exist" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.323512 4750 scope.go:117] "RemoveContainer" containerID="e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba" Mar 09 20:07:37 crc kubenswrapper[4750]: E0309 20:07:37.324042 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba\": container with ID starting with e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba not found: ID does not exist" containerID="e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.324106 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba"} err="failed to get container status \"e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba\": rpc error: code = NotFound desc = could not find container \"e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba\": container with ID starting with e54e1558e3c89328f86a320032e37501b4ab5f8898f412d30f1d5f656c317bba not found: ID does not exist" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.324139 4750 scope.go:117] "RemoveContainer" containerID="f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f" Mar 09 20:07:37 crc kubenswrapper[4750]: E0309 20:07:37.324734 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f\": container with ID starting with f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f not found: ID does not exist" containerID="f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.324767 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f"} err="failed to get container status \"f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f\": rpc error: code = NotFound desc = could not find container \"f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f\": container with ID starting with f0c302ca488bd95d47bd550a99c7f8b0931cd987d89be04e17ba6188dbdced8f not found: ID does not exist" Mar 09 20:07:37 crc kubenswrapper[4750]: I0309 20:07:37.387464 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" path="/var/lib/kubelet/pods/67d2b8fe-09ce-4c0b-9e56-998bf9d070a2/volumes" Mar 09 20:07:44 crc kubenswrapper[4750]: I0309 20:07:44.373984 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:07:44 crc kubenswrapper[4750]: E0309 20:07:44.374660 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:07:49 crc kubenswrapper[4750]: I0309 20:07:49.384556 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:07:49 crc kubenswrapper[4750]: E0309 20:07:49.386421 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:07:56 crc kubenswrapper[4750]: I0309 20:07:56.374438 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:07:56 crc kubenswrapper[4750]: E0309 20:07:56.376233 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.191072 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551448-ds2nd"] Mar 09 20:08:00 crc kubenswrapper[4750]: E0309 20:08:00.192460 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="extract-content" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.192475 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="extract-content" Mar 09 20:08:00 crc kubenswrapper[4750]: E0309 20:08:00.192498 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="extract-utilities" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.192505 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="extract-utilities" Mar 09 20:08:00 crc kubenswrapper[4750]: E0309 20:08:00.192542 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="registry-server" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.192549 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="registry-server" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.192774 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d2b8fe-09ce-4c0b-9e56-998bf9d070a2" containerName="registry-server" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.193539 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551448-ds2nd" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.197043 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.197335 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.197540 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.221219 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551448-ds2nd"] Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.295434 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2cs8\" (UniqueName: \"kubernetes.io/projected/ee4da78a-53ca-4ac8-b128-b943d6e58439-kube-api-access-l2cs8\") pod \"auto-csr-approver-29551448-ds2nd\" (UID: \"ee4da78a-53ca-4ac8-b128-b943d6e58439\") " pod="openshift-infra/auto-csr-approver-29551448-ds2nd" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.397142 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2cs8\" (UniqueName: \"kubernetes.io/projected/ee4da78a-53ca-4ac8-b128-b943d6e58439-kube-api-access-l2cs8\") pod \"auto-csr-approver-29551448-ds2nd\" (UID: \"ee4da78a-53ca-4ac8-b128-b943d6e58439\") " pod="openshift-infra/auto-csr-approver-29551448-ds2nd" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.417325 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2cs8\" (UniqueName: \"kubernetes.io/projected/ee4da78a-53ca-4ac8-b128-b943d6e58439-kube-api-access-l2cs8\") pod \"auto-csr-approver-29551448-ds2nd\" (UID: \"ee4da78a-53ca-4ac8-b128-b943d6e58439\") " pod="openshift-infra/auto-csr-approver-29551448-ds2nd" Mar 09 20:08:00 crc kubenswrapper[4750]: I0309 20:08:00.528274 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551448-ds2nd" Mar 09 20:08:01 crc kubenswrapper[4750]: I0309 20:08:01.006565 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551448-ds2nd"] Mar 09 20:08:01 crc kubenswrapper[4750]: I0309 20:08:01.449414 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551448-ds2nd" event={"ID":"ee4da78a-53ca-4ac8-b128-b943d6e58439","Type":"ContainerStarted","Data":"2718cf2359cd706f6ff06cccd0f179c26c3055f5a7e05c0b81ec250d0c61e398"} Mar 09 20:08:03 crc kubenswrapper[4750]: I0309 20:08:03.373182 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:08:03 crc kubenswrapper[4750]: E0309 20:08:03.373997 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:08:03 crc kubenswrapper[4750]: I0309 20:08:03.480319 4750 generic.go:334] "Generic (PLEG): container finished" podID="ee4da78a-53ca-4ac8-b128-b943d6e58439" containerID="e2caa00cfb6c067920b02dc8fda6cac663c433b844a2de3f5776f2db93503ec7" exitCode=0 Mar 09 20:08:03 crc kubenswrapper[4750]: I0309 20:08:03.480369 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551448-ds2nd" event={"ID":"ee4da78a-53ca-4ac8-b128-b943d6e58439","Type":"ContainerDied","Data":"e2caa00cfb6c067920b02dc8fda6cac663c433b844a2de3f5776f2db93503ec7"} Mar 09 20:08:04 crc kubenswrapper[4750]: I0309 20:08:04.869610 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551448-ds2nd" Mar 09 20:08:04 crc kubenswrapper[4750]: I0309 20:08:04.992843 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2cs8\" (UniqueName: \"kubernetes.io/projected/ee4da78a-53ca-4ac8-b128-b943d6e58439-kube-api-access-l2cs8\") pod \"ee4da78a-53ca-4ac8-b128-b943d6e58439\" (UID: \"ee4da78a-53ca-4ac8-b128-b943d6e58439\") " Mar 09 20:08:05 crc kubenswrapper[4750]: I0309 20:08:05.001564 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee4da78a-53ca-4ac8-b128-b943d6e58439-kube-api-access-l2cs8" (OuterVolumeSpecName: "kube-api-access-l2cs8") pod "ee4da78a-53ca-4ac8-b128-b943d6e58439" (UID: "ee4da78a-53ca-4ac8-b128-b943d6e58439"). InnerVolumeSpecName "kube-api-access-l2cs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:08:05 crc kubenswrapper[4750]: I0309 20:08:05.095244 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2cs8\" (UniqueName: \"kubernetes.io/projected/ee4da78a-53ca-4ac8-b128-b943d6e58439-kube-api-access-l2cs8\") on node \"crc\" DevicePath \"\"" Mar 09 20:08:05 crc kubenswrapper[4750]: I0309 20:08:05.502671 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551448-ds2nd" event={"ID":"ee4da78a-53ca-4ac8-b128-b943d6e58439","Type":"ContainerDied","Data":"2718cf2359cd706f6ff06cccd0f179c26c3055f5a7e05c0b81ec250d0c61e398"} Mar 09 20:08:05 crc kubenswrapper[4750]: I0309 20:08:05.502721 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2718cf2359cd706f6ff06cccd0f179c26c3055f5a7e05c0b81ec250d0c61e398" Mar 09 20:08:05 crc kubenswrapper[4750]: I0309 20:08:05.502743 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551448-ds2nd" Mar 09 20:08:05 crc kubenswrapper[4750]: I0309 20:08:05.959211 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551442-ldtvn"] Mar 09 20:08:05 crc kubenswrapper[4750]: I0309 20:08:05.968565 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551442-ldtvn"] Mar 09 20:08:07 crc kubenswrapper[4750]: I0309 20:08:07.374772 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:08:07 crc kubenswrapper[4750]: E0309 20:08:07.375369 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:08:07 crc kubenswrapper[4750]: I0309 20:08:07.384738 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c8f040d-1d00-42c3-94e5-bec6babbbe48" path="/var/lib/kubelet/pods/5c8f040d-1d00-42c3-94e5-bec6babbbe48/volumes" Mar 09 20:08:14 crc kubenswrapper[4750]: I0309 20:08:14.373897 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:08:14 crc kubenswrapper[4750]: E0309 20:08:14.374836 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:08:21 crc kubenswrapper[4750]: I0309 20:08:21.374415 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:08:21 crc kubenswrapper[4750]: E0309 20:08:21.375413 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:08:28 crc kubenswrapper[4750]: I0309 20:08:28.374374 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:08:28 crc kubenswrapper[4750]: E0309 20:08:28.375330 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:08:36 crc kubenswrapper[4750]: I0309 20:08:36.373458 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:08:36 crc kubenswrapper[4750]: E0309 20:08:36.374395 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:08:42 crc kubenswrapper[4750]: I0309 20:08:42.374515 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:08:42 crc kubenswrapper[4750]: E0309 20:08:42.376737 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:08:50 crc kubenswrapper[4750]: I0309 20:08:50.374081 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:08:50 crc kubenswrapper[4750]: E0309 20:08:50.374811 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:08:52 crc kubenswrapper[4750]: I0309 20:08:52.373563 4750 scope.go:117] "RemoveContainer" containerID="a361870b2a77fa92467d58b17a999624ac65935640121eec7535ff3f005d72dc" Mar 09 20:08:53 crc kubenswrapper[4750]: I0309 20:08:53.374338 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:08:53 crc kubenswrapper[4750]: E0309 20:08:53.374682 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:09:05 crc kubenswrapper[4750]: I0309 20:09:05.373713 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:09:05 crc kubenswrapper[4750]: E0309 20:09:05.374731 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:09:07 crc kubenswrapper[4750]: I0309 20:09:07.373943 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:09:07 crc kubenswrapper[4750]: E0309 20:09:07.374529 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:09:17 crc kubenswrapper[4750]: I0309 20:09:17.373276 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:09:17 crc kubenswrapper[4750]: E0309 20:09:17.374088 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:09:18 crc kubenswrapper[4750]: I0309 20:09:18.374290 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:09:18 crc kubenswrapper[4750]: E0309 20:09:18.374716 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:09:30 crc kubenswrapper[4750]: I0309 20:09:30.375047 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:09:30 crc kubenswrapper[4750]: E0309 20:09:30.377224 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:09:31 crc kubenswrapper[4750]: I0309 20:09:31.375062 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:09:31 crc kubenswrapper[4750]: E0309 20:09:31.375480 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:09:45 crc kubenswrapper[4750]: I0309 20:09:45.373799 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:09:45 crc kubenswrapper[4750]: E0309 20:09:45.374491 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:09:46 crc kubenswrapper[4750]: I0309 20:09:46.374611 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:09:46 crc kubenswrapper[4750]: E0309 20:09:46.375383 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:09:58 crc kubenswrapper[4750]: I0309 20:09:58.373736 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:09:58 crc kubenswrapper[4750]: E0309 20:09:58.374528 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.143283 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551450-zn4xs"] Mar 09 20:10:00 crc kubenswrapper[4750]: E0309 20:10:00.143916 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4da78a-53ca-4ac8-b128-b943d6e58439" containerName="oc" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.143928 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4da78a-53ca-4ac8-b128-b943d6e58439" containerName="oc" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.144129 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4da78a-53ca-4ac8-b128-b943d6e58439" containerName="oc" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.144754 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551450-zn4xs" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.146360 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.147184 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.148192 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.167325 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551450-zn4xs"] Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.286181 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kblhw\" (UniqueName: \"kubernetes.io/projected/85c695cb-27f6-4a1d-9065-798fff27f528-kube-api-access-kblhw\") pod \"auto-csr-approver-29551450-zn4xs\" (UID: \"85c695cb-27f6-4a1d-9065-798fff27f528\") " pod="openshift-infra/auto-csr-approver-29551450-zn4xs" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.373264 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:10:00 crc kubenswrapper[4750]: E0309 20:10:00.373556 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.388013 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kblhw\" (UniqueName: \"kubernetes.io/projected/85c695cb-27f6-4a1d-9065-798fff27f528-kube-api-access-kblhw\") pod \"auto-csr-approver-29551450-zn4xs\" (UID: \"85c695cb-27f6-4a1d-9065-798fff27f528\") " pod="openshift-infra/auto-csr-approver-29551450-zn4xs" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.408455 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kblhw\" (UniqueName: \"kubernetes.io/projected/85c695cb-27f6-4a1d-9065-798fff27f528-kube-api-access-kblhw\") pod \"auto-csr-approver-29551450-zn4xs\" (UID: \"85c695cb-27f6-4a1d-9065-798fff27f528\") " pod="openshift-infra/auto-csr-approver-29551450-zn4xs" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.466998 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551450-zn4xs" Mar 09 20:10:00 crc kubenswrapper[4750]: I0309 20:10:00.965707 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551450-zn4xs"] Mar 09 20:10:01 crc kubenswrapper[4750]: I0309 20:10:01.723807 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551450-zn4xs" event={"ID":"85c695cb-27f6-4a1d-9065-798fff27f528","Type":"ContainerStarted","Data":"738e9194ac903a5271af1e41bc8179b1b6f927fe23ebcf4e2577acf2a38bb051"} Mar 09 20:10:03 crc kubenswrapper[4750]: I0309 20:10:03.743491 4750 generic.go:334] "Generic (PLEG): container finished" podID="85c695cb-27f6-4a1d-9065-798fff27f528" containerID="553e65245cc8b523056bc0fd43895fc062ed9035198aa566a3c729ae1daecaf8" exitCode=0 Mar 09 20:10:03 crc kubenswrapper[4750]: I0309 20:10:03.743593 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551450-zn4xs" event={"ID":"85c695cb-27f6-4a1d-9065-798fff27f528","Type":"ContainerDied","Data":"553e65245cc8b523056bc0fd43895fc062ed9035198aa566a3c729ae1daecaf8"} Mar 09 20:10:05 crc kubenswrapper[4750]: I0309 20:10:05.094843 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551450-zn4xs" Mar 09 20:10:05 crc kubenswrapper[4750]: I0309 20:10:05.193302 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kblhw\" (UniqueName: \"kubernetes.io/projected/85c695cb-27f6-4a1d-9065-798fff27f528-kube-api-access-kblhw\") pod \"85c695cb-27f6-4a1d-9065-798fff27f528\" (UID: \"85c695cb-27f6-4a1d-9065-798fff27f528\") " Mar 09 20:10:05 crc kubenswrapper[4750]: I0309 20:10:05.200163 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c695cb-27f6-4a1d-9065-798fff27f528-kube-api-access-kblhw" (OuterVolumeSpecName: "kube-api-access-kblhw") pod "85c695cb-27f6-4a1d-9065-798fff27f528" (UID: "85c695cb-27f6-4a1d-9065-798fff27f528"). InnerVolumeSpecName "kube-api-access-kblhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:10:05 crc kubenswrapper[4750]: I0309 20:10:05.296610 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kblhw\" (UniqueName: \"kubernetes.io/projected/85c695cb-27f6-4a1d-9065-798fff27f528-kube-api-access-kblhw\") on node \"crc\" DevicePath \"\"" Mar 09 20:10:05 crc kubenswrapper[4750]: I0309 20:10:05.764270 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551450-zn4xs" event={"ID":"85c695cb-27f6-4a1d-9065-798fff27f528","Type":"ContainerDied","Data":"738e9194ac903a5271af1e41bc8179b1b6f927fe23ebcf4e2577acf2a38bb051"} Mar 09 20:10:05 crc kubenswrapper[4750]: I0309 20:10:05.764319 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="738e9194ac903a5271af1e41bc8179b1b6f927fe23ebcf4e2577acf2a38bb051" Mar 09 20:10:05 crc kubenswrapper[4750]: I0309 20:10:05.764366 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551450-zn4xs" Mar 09 20:10:06 crc kubenswrapper[4750]: I0309 20:10:06.165979 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551444-stpgq"] Mar 09 20:10:06 crc kubenswrapper[4750]: I0309 20:10:06.176150 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551444-stpgq"] Mar 09 20:10:07 crc kubenswrapper[4750]: I0309 20:10:07.384877 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e704c07-6189-469e-87fd-10f711cbc618" path="/var/lib/kubelet/pods/3e704c07-6189-469e-87fd-10f711cbc618/volumes" Mar 09 20:10:12 crc kubenswrapper[4750]: I0309 20:10:12.374182 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:10:12 crc kubenswrapper[4750]: E0309 20:10:12.375719 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:10:13 crc kubenswrapper[4750]: I0309 20:10:13.374621 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:10:13 crc kubenswrapper[4750]: E0309 20:10:13.375293 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:10:23 crc kubenswrapper[4750]: I0309 20:10:23.373533 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:10:23 crc kubenswrapper[4750]: E0309 20:10:23.374326 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:10:27 crc kubenswrapper[4750]: I0309 20:10:27.374249 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:10:27 crc kubenswrapper[4750]: E0309 20:10:27.375563 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:10:34 crc kubenswrapper[4750]: I0309 20:10:34.374112 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:10:34 crc kubenswrapper[4750]: E0309 20:10:34.375279 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:10:40 crc kubenswrapper[4750]: I0309 20:10:40.374711 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:10:40 crc kubenswrapper[4750]: E0309 20:10:40.376308 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:10:45 crc kubenswrapper[4750]: I0309 20:10:45.374774 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:10:45 crc kubenswrapper[4750]: E0309 20:10:45.375710 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:10:52 crc kubenswrapper[4750]: I0309 20:10:52.520375 4750 scope.go:117] "RemoveContainer" containerID="0e91a60a9bdded50c0d7315cfe08b58aa8ac24cb8adb4b758cb4b553d18a4626" Mar 09 20:10:54 crc kubenswrapper[4750]: I0309 20:10:54.373295 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:10:54 crc kubenswrapper[4750]: E0309 20:10:54.374136 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:11:00 crc kubenswrapper[4750]: I0309 20:11:00.374345 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:11:00 crc kubenswrapper[4750]: E0309 20:11:00.376318 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:11:09 crc kubenswrapper[4750]: I0309 20:11:09.381730 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:11:09 crc kubenswrapper[4750]: E0309 20:11:09.382582 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:11:14 crc kubenswrapper[4750]: I0309 20:11:14.373424 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:11:14 crc kubenswrapper[4750]: E0309 20:11:14.374373 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.248746 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7b7rn"] Mar 09 20:11:15 crc kubenswrapper[4750]: E0309 20:11:15.249569 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c695cb-27f6-4a1d-9065-798fff27f528" containerName="oc" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.249589 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c695cb-27f6-4a1d-9065-798fff27f528" containerName="oc" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.249841 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c695cb-27f6-4a1d-9065-798fff27f528" containerName="oc" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.251852 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.259585 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7b7rn"] Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.286609 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-utilities\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.286702 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr5wf\" (UniqueName: \"kubernetes.io/projected/13850ffe-fd6f-4ca7-a264-672110a58e6b-kube-api-access-pr5wf\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.286739 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-catalog-content\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.390004 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-utilities\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.390061 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr5wf\" (UniqueName: \"kubernetes.io/projected/13850ffe-fd6f-4ca7-a264-672110a58e6b-kube-api-access-pr5wf\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.390090 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-catalog-content\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.390936 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-utilities\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.390955 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-catalog-content\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.409547 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr5wf\" (UniqueName: \"kubernetes.io/projected/13850ffe-fd6f-4ca7-a264-672110a58e6b-kube-api-access-pr5wf\") pod \"certified-operators-7b7rn\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:15 crc kubenswrapper[4750]: I0309 20:11:15.581699 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:16 crc kubenswrapper[4750]: I0309 20:11:16.093570 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7b7rn"] Mar 09 20:11:16 crc kubenswrapper[4750]: I0309 20:11:16.504556 4750 generic.go:334] "Generic (PLEG): container finished" podID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerID="9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c" exitCode=0 Mar 09 20:11:16 crc kubenswrapper[4750]: I0309 20:11:16.504619 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7b7rn" event={"ID":"13850ffe-fd6f-4ca7-a264-672110a58e6b","Type":"ContainerDied","Data":"9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c"} Mar 09 20:11:16 crc kubenswrapper[4750]: I0309 20:11:16.504845 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7b7rn" event={"ID":"13850ffe-fd6f-4ca7-a264-672110a58e6b","Type":"ContainerStarted","Data":"32050b6000bebaa9e255182671852b90add695b07bd095ab9c5105ac0b8c041c"} Mar 09 20:11:17 crc kubenswrapper[4750]: I0309 20:11:17.520496 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7b7rn" event={"ID":"13850ffe-fd6f-4ca7-a264-672110a58e6b","Type":"ContainerStarted","Data":"a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425"} Mar 09 20:11:18 crc kubenswrapper[4750]: I0309 20:11:18.534031 4750 generic.go:334] "Generic (PLEG): container finished" podID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerID="a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425" exitCode=0 Mar 09 20:11:18 crc kubenswrapper[4750]: I0309 20:11:18.534148 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7b7rn" event={"ID":"13850ffe-fd6f-4ca7-a264-672110a58e6b","Type":"ContainerDied","Data":"a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425"} Mar 09 20:11:19 crc kubenswrapper[4750]: I0309 20:11:19.555347 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7b7rn" event={"ID":"13850ffe-fd6f-4ca7-a264-672110a58e6b","Type":"ContainerStarted","Data":"4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048"} Mar 09 20:11:19 crc kubenswrapper[4750]: I0309 20:11:19.578184 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7b7rn" podStartSLOduration=2.083978108 podStartE2EDuration="4.57816643s" podCreationTimestamp="2026-03-09 20:11:15 +0000 UTC" firstStartedPulling="2026-03-09 20:11:16.50613263 +0000 UTC m=+6357.848605028" lastFinishedPulling="2026-03-09 20:11:19.000320942 +0000 UTC m=+6360.342793350" observedRunningTime="2026-03-09 20:11:19.575156598 +0000 UTC m=+6360.917628996" watchObservedRunningTime="2026-03-09 20:11:19.57816643 +0000 UTC m=+6360.920638828" Mar 09 20:11:20 crc kubenswrapper[4750]: I0309 20:11:20.373577 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:11:20 crc kubenswrapper[4750]: E0309 20:11:20.374096 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:11:25 crc kubenswrapper[4750]: I0309 20:11:25.582984 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:25 crc kubenswrapper[4750]: I0309 20:11:25.583581 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:25 crc kubenswrapper[4750]: I0309 20:11:25.629382 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:25 crc kubenswrapper[4750]: I0309 20:11:25.683083 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:25 crc kubenswrapper[4750]: I0309 20:11:25.865865 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7b7rn"] Mar 09 20:11:27 crc kubenswrapper[4750]: I0309 20:11:27.631028 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7b7rn" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="registry-server" containerID="cri-o://4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048" gracePeriod=2 Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.196058 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.374031 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.390972 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-catalog-content\") pod \"13850ffe-fd6f-4ca7-a264-672110a58e6b\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.391273 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr5wf\" (UniqueName: \"kubernetes.io/projected/13850ffe-fd6f-4ca7-a264-672110a58e6b-kube-api-access-pr5wf\") pod \"13850ffe-fd6f-4ca7-a264-672110a58e6b\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.391422 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-utilities\") pod \"13850ffe-fd6f-4ca7-a264-672110a58e6b\" (UID: \"13850ffe-fd6f-4ca7-a264-672110a58e6b\") " Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.392870 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-utilities" (OuterVolumeSpecName: "utilities") pod "13850ffe-fd6f-4ca7-a264-672110a58e6b" (UID: "13850ffe-fd6f-4ca7-a264-672110a58e6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.398326 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13850ffe-fd6f-4ca7-a264-672110a58e6b-kube-api-access-pr5wf" (OuterVolumeSpecName: "kube-api-access-pr5wf") pod "13850ffe-fd6f-4ca7-a264-672110a58e6b" (UID: "13850ffe-fd6f-4ca7-a264-672110a58e6b"). InnerVolumeSpecName "kube-api-access-pr5wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.448411 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13850ffe-fd6f-4ca7-a264-672110a58e6b" (UID: "13850ffe-fd6f-4ca7-a264-672110a58e6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.495725 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.496051 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr5wf\" (UniqueName: \"kubernetes.io/projected/13850ffe-fd6f-4ca7-a264-672110a58e6b-kube-api-access-pr5wf\") on node \"crc\" DevicePath \"\"" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.496069 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13850ffe-fd6f-4ca7-a264-672110a58e6b-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.641991 4750 generic.go:334] "Generic (PLEG): container finished" podID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerID="4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048" exitCode=0 Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.642039 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7b7rn" event={"ID":"13850ffe-fd6f-4ca7-a264-672110a58e6b","Type":"ContainerDied","Data":"4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048"} Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.642055 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7b7rn" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.642069 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7b7rn" event={"ID":"13850ffe-fd6f-4ca7-a264-672110a58e6b","Type":"ContainerDied","Data":"32050b6000bebaa9e255182671852b90add695b07bd095ab9c5105ac0b8c041c"} Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.642087 4750 scope.go:117] "RemoveContainer" containerID="4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.685747 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7b7rn"] Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.687586 4750 scope.go:117] "RemoveContainer" containerID="a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.698001 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7b7rn"] Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.719448 4750 scope.go:117] "RemoveContainer" containerID="9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.743762 4750 scope.go:117] "RemoveContainer" containerID="4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048" Mar 09 20:11:28 crc kubenswrapper[4750]: E0309 20:11:28.752011 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048\": container with ID starting with 4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048 not found: ID does not exist" containerID="4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.752068 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048"} err="failed to get container status \"4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048\": rpc error: code = NotFound desc = could not find container \"4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048\": container with ID starting with 4d78e1918ef32d1e45d37722999373f0387120d59a0de669211c7d1ba9a12048 not found: ID does not exist" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.752093 4750 scope.go:117] "RemoveContainer" containerID="a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425" Mar 09 20:11:28 crc kubenswrapper[4750]: E0309 20:11:28.752473 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425\": container with ID starting with a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425 not found: ID does not exist" containerID="a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.752519 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425"} err="failed to get container status \"a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425\": rpc error: code = NotFound desc = could not find container \"a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425\": container with ID starting with a3a376747fb5625bed35275b03390145a6c934648d578d8c8467521a2ebbc425 not found: ID does not exist" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.752536 4750 scope.go:117] "RemoveContainer" containerID="9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c" Mar 09 20:11:28 crc kubenswrapper[4750]: E0309 20:11:28.752720 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c\": container with ID starting with 9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c not found: ID does not exist" containerID="9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c" Mar 09 20:11:28 crc kubenswrapper[4750]: I0309 20:11:28.752735 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c"} err="failed to get container status \"9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c\": rpc error: code = NotFound desc = could not find container \"9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c\": container with ID starting with 9b1dc2cb64eae8369d6945d7907b21dfd1ebd816c5c371abe943348f586b570c not found: ID does not exist" Mar 09 20:11:29 crc kubenswrapper[4750]: I0309 20:11:29.391141 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" path="/var/lib/kubelet/pods/13850ffe-fd6f-4ca7-a264-672110a58e6b/volumes" Mar 09 20:11:29 crc kubenswrapper[4750]: I0309 20:11:29.675680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375"} Mar 09 20:11:30 crc kubenswrapper[4750]: I0309 20:11:30.190078 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:11:31 crc kubenswrapper[4750]: I0309 20:11:31.702369 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" exitCode=1 Mar 09 20:11:31 crc kubenswrapper[4750]: I0309 20:11:31.702440 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375"} Mar 09 20:11:31 crc kubenswrapper[4750]: I0309 20:11:31.702657 4750 scope.go:117] "RemoveContainer" containerID="e97c3157722cd1e864008fefb43d535e5469c601cef10be6ff8bbf76a94fb730" Mar 09 20:11:31 crc kubenswrapper[4750]: I0309 20:11:31.703855 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:11:31 crc kubenswrapper[4750]: E0309 20:11:31.704599 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:11:32 crc kubenswrapper[4750]: I0309 20:11:32.191722 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:11:32 crc kubenswrapper[4750]: I0309 20:11:32.718222 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:11:32 crc kubenswrapper[4750]: E0309 20:11:32.718683 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:11:33 crc kubenswrapper[4750]: I0309 20:11:33.374470 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:11:33 crc kubenswrapper[4750]: I0309 20:11:33.732922 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"1aee94e71f7a10634c557910d5c09ce69ff26474679f02e5aab3816f75d7901e"} Mar 09 20:11:35 crc kubenswrapper[4750]: I0309 20:11:35.190040 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:11:35 crc kubenswrapper[4750]: I0309 20:11:35.191171 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:11:35 crc kubenswrapper[4750]: E0309 20:11:35.191520 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:11:48 crc kubenswrapper[4750]: I0309 20:11:48.374201 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:11:48 crc kubenswrapper[4750]: E0309 20:11:48.375308 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:11:59 crc kubenswrapper[4750]: I0309 20:11:59.380397 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:11:59 crc kubenswrapper[4750]: E0309 20:11:59.381815 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.174763 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551452-nt87s"] Mar 09 20:12:00 crc kubenswrapper[4750]: E0309 20:12:00.175613 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="extract-utilities" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.175695 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="extract-utilities" Mar 09 20:12:00 crc kubenswrapper[4750]: E0309 20:12:00.175731 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="registry-server" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.175776 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="registry-server" Mar 09 20:12:00 crc kubenswrapper[4750]: E0309 20:12:00.175830 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="extract-content" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.175849 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="extract-content" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.176403 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="13850ffe-fd6f-4ca7-a264-672110a58e6b" containerName="registry-server" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.177873 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551452-nt87s" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.185006 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.185657 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.187583 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.199212 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551452-nt87s"] Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.307121 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvwmx\" (UniqueName: \"kubernetes.io/projected/c136bd51-ed7b-4b3b-988e-420eadff782c-kube-api-access-bvwmx\") pod \"auto-csr-approver-29551452-nt87s\" (UID: \"c136bd51-ed7b-4b3b-988e-420eadff782c\") " pod="openshift-infra/auto-csr-approver-29551452-nt87s" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.410320 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvwmx\" (UniqueName: \"kubernetes.io/projected/c136bd51-ed7b-4b3b-988e-420eadff782c-kube-api-access-bvwmx\") pod \"auto-csr-approver-29551452-nt87s\" (UID: \"c136bd51-ed7b-4b3b-988e-420eadff782c\") " pod="openshift-infra/auto-csr-approver-29551452-nt87s" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.446518 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvwmx\" (UniqueName: \"kubernetes.io/projected/c136bd51-ed7b-4b3b-988e-420eadff782c-kube-api-access-bvwmx\") pod \"auto-csr-approver-29551452-nt87s\" (UID: \"c136bd51-ed7b-4b3b-988e-420eadff782c\") " pod="openshift-infra/auto-csr-approver-29551452-nt87s" Mar 09 20:12:00 crc kubenswrapper[4750]: I0309 20:12:00.514714 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551452-nt87s" Mar 09 20:12:01 crc kubenswrapper[4750]: I0309 20:12:01.035309 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551452-nt87s"] Mar 09 20:12:01 crc kubenswrapper[4750]: W0309 20:12:01.045166 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc136bd51_ed7b_4b3b_988e_420eadff782c.slice/crio-d43e0524f65096230195f71990079c4e1a83e9427db9e3b12679aef9c8f2fcda WatchSource:0}: Error finding container d43e0524f65096230195f71990079c4e1a83e9427db9e3b12679aef9c8f2fcda: Status 404 returned error can't find the container with id d43e0524f65096230195f71990079c4e1a83e9427db9e3b12679aef9c8f2fcda Mar 09 20:12:02 crc kubenswrapper[4750]: I0309 20:12:02.055866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551452-nt87s" event={"ID":"c136bd51-ed7b-4b3b-988e-420eadff782c","Type":"ContainerStarted","Data":"d43e0524f65096230195f71990079c4e1a83e9427db9e3b12679aef9c8f2fcda"} Mar 09 20:12:04 crc kubenswrapper[4750]: I0309 20:12:04.074023 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551452-nt87s" event={"ID":"c136bd51-ed7b-4b3b-988e-420eadff782c","Type":"ContainerStarted","Data":"2a4e4e4d9410d1c6f872f7e1b34230ab9bf220fe59359782563ffef9410d2271"} Mar 09 20:12:04 crc kubenswrapper[4750]: I0309 20:12:04.089917 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29551452-nt87s" podStartSLOduration=1.414282002 podStartE2EDuration="4.089900186s" podCreationTimestamp="2026-03-09 20:12:00 +0000 UTC" firstStartedPulling="2026-03-09 20:12:01.048387584 +0000 UTC m=+6402.390859982" lastFinishedPulling="2026-03-09 20:12:03.724005768 +0000 UTC m=+6405.066478166" observedRunningTime="2026-03-09 20:12:04.087894042 +0000 UTC m=+6405.430366440" watchObservedRunningTime="2026-03-09 20:12:04.089900186 +0000 UTC m=+6405.432372584" Mar 09 20:12:05 crc kubenswrapper[4750]: I0309 20:12:05.085674 4750 generic.go:334] "Generic (PLEG): container finished" podID="c136bd51-ed7b-4b3b-988e-420eadff782c" containerID="2a4e4e4d9410d1c6f872f7e1b34230ab9bf220fe59359782563ffef9410d2271" exitCode=0 Mar 09 20:12:05 crc kubenswrapper[4750]: I0309 20:12:05.085797 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551452-nt87s" event={"ID":"c136bd51-ed7b-4b3b-988e-420eadff782c","Type":"ContainerDied","Data":"2a4e4e4d9410d1c6f872f7e1b34230ab9bf220fe59359782563ffef9410d2271"} Mar 09 20:12:06 crc kubenswrapper[4750]: I0309 20:12:06.437391 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551452-nt87s" Mar 09 20:12:06 crc kubenswrapper[4750]: I0309 20:12:06.535720 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvwmx\" (UniqueName: \"kubernetes.io/projected/c136bd51-ed7b-4b3b-988e-420eadff782c-kube-api-access-bvwmx\") pod \"c136bd51-ed7b-4b3b-988e-420eadff782c\" (UID: \"c136bd51-ed7b-4b3b-988e-420eadff782c\") " Mar 09 20:12:06 crc kubenswrapper[4750]: I0309 20:12:06.541819 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c136bd51-ed7b-4b3b-988e-420eadff782c-kube-api-access-bvwmx" (OuterVolumeSpecName: "kube-api-access-bvwmx") pod "c136bd51-ed7b-4b3b-988e-420eadff782c" (UID: "c136bd51-ed7b-4b3b-988e-420eadff782c"). InnerVolumeSpecName "kube-api-access-bvwmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:12:06 crc kubenswrapper[4750]: I0309 20:12:06.638519 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvwmx\" (UniqueName: \"kubernetes.io/projected/c136bd51-ed7b-4b3b-988e-420eadff782c-kube-api-access-bvwmx\") on node \"crc\" DevicePath \"\"" Mar 09 20:12:07 crc kubenswrapper[4750]: I0309 20:12:07.107416 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551452-nt87s" event={"ID":"c136bd51-ed7b-4b3b-988e-420eadff782c","Type":"ContainerDied","Data":"d43e0524f65096230195f71990079c4e1a83e9427db9e3b12679aef9c8f2fcda"} Mar 09 20:12:07 crc kubenswrapper[4750]: I0309 20:12:07.107778 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d43e0524f65096230195f71990079c4e1a83e9427db9e3b12679aef9c8f2fcda" Mar 09 20:12:07 crc kubenswrapper[4750]: I0309 20:12:07.107455 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551452-nt87s" Mar 09 20:12:07 crc kubenswrapper[4750]: I0309 20:12:07.195594 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551446-cqtrz"] Mar 09 20:12:07 crc kubenswrapper[4750]: I0309 20:12:07.207777 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551446-cqtrz"] Mar 09 20:12:07 crc kubenswrapper[4750]: I0309 20:12:07.386470 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f1ae607-f58d-4b0f-a484-b7c40889adde" path="/var/lib/kubelet/pods/2f1ae607-f58d-4b0f-a484-b7c40889adde/volumes" Mar 09 20:12:11 crc kubenswrapper[4750]: I0309 20:12:11.373980 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:12:11 crc kubenswrapper[4750]: E0309 20:12:11.375191 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.337437 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-97bq5"] Mar 09 20:12:13 crc kubenswrapper[4750]: E0309 20:12:13.338342 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c136bd51-ed7b-4b3b-988e-420eadff782c" containerName="oc" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.338366 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c136bd51-ed7b-4b3b-988e-420eadff782c" containerName="oc" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.338678 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c136bd51-ed7b-4b3b-988e-420eadff782c" containerName="oc" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.341533 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.368048 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97bq5"] Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.393314 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-catalog-content\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.393429 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-utilities\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.393461 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4pxd\" (UniqueName: \"kubernetes.io/projected/7522e5a9-274a-4a16-b508-8418e65bebaf-kube-api-access-b4pxd\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.494835 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-catalog-content\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.494938 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-utilities\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.494959 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4pxd\" (UniqueName: \"kubernetes.io/projected/7522e5a9-274a-4a16-b508-8418e65bebaf-kube-api-access-b4pxd\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.495364 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-catalog-content\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.495441 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-utilities\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.527423 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4pxd\" (UniqueName: \"kubernetes.io/projected/7522e5a9-274a-4a16-b508-8418e65bebaf-kube-api-access-b4pxd\") pod \"community-operators-97bq5\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:13 crc kubenswrapper[4750]: I0309 20:12:13.663298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:14 crc kubenswrapper[4750]: I0309 20:12:14.332405 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97bq5"] Mar 09 20:12:15 crc kubenswrapper[4750]: I0309 20:12:15.184767 4750 generic.go:334] "Generic (PLEG): container finished" podID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerID="22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c" exitCode=0 Mar 09 20:12:15 crc kubenswrapper[4750]: I0309 20:12:15.184841 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97bq5" event={"ID":"7522e5a9-274a-4a16-b508-8418e65bebaf","Type":"ContainerDied","Data":"22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c"} Mar 09 20:12:15 crc kubenswrapper[4750]: I0309 20:12:15.185153 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97bq5" event={"ID":"7522e5a9-274a-4a16-b508-8418e65bebaf","Type":"ContainerStarted","Data":"5a291d6c090b0c82c65f9b0ecc21ac9f3d50f84782539d56cce7c2658f2d47e7"} Mar 09 20:12:17 crc kubenswrapper[4750]: I0309 20:12:17.209096 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97bq5" event={"ID":"7522e5a9-274a-4a16-b508-8418e65bebaf","Type":"ContainerStarted","Data":"4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b"} Mar 09 20:12:18 crc kubenswrapper[4750]: I0309 20:12:18.232955 4750 generic.go:334] "Generic (PLEG): container finished" podID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerID="4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b" exitCode=0 Mar 09 20:12:18 crc kubenswrapper[4750]: I0309 20:12:18.233078 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97bq5" event={"ID":"7522e5a9-274a-4a16-b508-8418e65bebaf","Type":"ContainerDied","Data":"4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b"} Mar 09 20:12:19 crc kubenswrapper[4750]: I0309 20:12:19.253045 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97bq5" event={"ID":"7522e5a9-274a-4a16-b508-8418e65bebaf","Type":"ContainerStarted","Data":"878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34"} Mar 09 20:12:19 crc kubenswrapper[4750]: I0309 20:12:19.295863 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-97bq5" podStartSLOduration=2.800300229 podStartE2EDuration="6.295834156s" podCreationTimestamp="2026-03-09 20:12:13 +0000 UTC" firstStartedPulling="2026-03-09 20:12:15.187210967 +0000 UTC m=+6416.529683415" lastFinishedPulling="2026-03-09 20:12:18.682744944 +0000 UTC m=+6420.025217342" observedRunningTime="2026-03-09 20:12:19.284042077 +0000 UTC m=+6420.626514475" watchObservedRunningTime="2026-03-09 20:12:19.295834156 +0000 UTC m=+6420.638306574" Mar 09 20:12:22 crc kubenswrapper[4750]: I0309 20:12:22.374286 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:12:22 crc kubenswrapper[4750]: E0309 20:12:22.375157 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:12:23 crc kubenswrapper[4750]: I0309 20:12:23.664007 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:23 crc kubenswrapper[4750]: I0309 20:12:23.664284 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:23 crc kubenswrapper[4750]: I0309 20:12:23.722036 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:24 crc kubenswrapper[4750]: I0309 20:12:24.359686 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:24 crc kubenswrapper[4750]: I0309 20:12:24.423708 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97bq5"] Mar 09 20:12:26 crc kubenswrapper[4750]: I0309 20:12:26.324316 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-97bq5" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="registry-server" containerID="cri-o://878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34" gracePeriod=2 Mar 09 20:12:26 crc kubenswrapper[4750]: I0309 20:12:26.945617 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.023126 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4pxd\" (UniqueName: \"kubernetes.io/projected/7522e5a9-274a-4a16-b508-8418e65bebaf-kube-api-access-b4pxd\") pod \"7522e5a9-274a-4a16-b508-8418e65bebaf\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.023258 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-utilities\") pod \"7522e5a9-274a-4a16-b508-8418e65bebaf\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.023351 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-catalog-content\") pod \"7522e5a9-274a-4a16-b508-8418e65bebaf\" (UID: \"7522e5a9-274a-4a16-b508-8418e65bebaf\") " Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.024093 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-utilities" (OuterVolumeSpecName: "utilities") pod "7522e5a9-274a-4a16-b508-8418e65bebaf" (UID: "7522e5a9-274a-4a16-b508-8418e65bebaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.032910 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7522e5a9-274a-4a16-b508-8418e65bebaf-kube-api-access-b4pxd" (OuterVolumeSpecName: "kube-api-access-b4pxd") pod "7522e5a9-274a-4a16-b508-8418e65bebaf" (UID: "7522e5a9-274a-4a16-b508-8418e65bebaf"). InnerVolumeSpecName "kube-api-access-b4pxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.096157 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7522e5a9-274a-4a16-b508-8418e65bebaf" (UID: "7522e5a9-274a-4a16-b508-8418e65bebaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.125503 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4pxd\" (UniqueName: \"kubernetes.io/projected/7522e5a9-274a-4a16-b508-8418e65bebaf-kube-api-access-b4pxd\") on node \"crc\" DevicePath \"\"" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.125548 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.125561 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7522e5a9-274a-4a16-b508-8418e65bebaf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.336039 4750 generic.go:334] "Generic (PLEG): container finished" podID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerID="878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34" exitCode=0 Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.336129 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97bq5" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.336151 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97bq5" event={"ID":"7522e5a9-274a-4a16-b508-8418e65bebaf","Type":"ContainerDied","Data":"878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34"} Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.336991 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97bq5" event={"ID":"7522e5a9-274a-4a16-b508-8418e65bebaf","Type":"ContainerDied","Data":"5a291d6c090b0c82c65f9b0ecc21ac9f3d50f84782539d56cce7c2658f2d47e7"} Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.337011 4750 scope.go:117] "RemoveContainer" containerID="878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.365559 4750 scope.go:117] "RemoveContainer" containerID="4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.436973 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97bq5"] Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.437018 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-97bq5"] Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.437171 4750 scope.go:117] "RemoveContainer" containerID="22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.470283 4750 scope.go:117] "RemoveContainer" containerID="878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34" Mar 09 20:12:27 crc kubenswrapper[4750]: E0309 20:12:27.470875 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34\": container with ID starting with 878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34 not found: ID does not exist" containerID="878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.472678 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34"} err="failed to get container status \"878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34\": rpc error: code = NotFound desc = could not find container \"878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34\": container with ID starting with 878248b64877ba41e6f9b1f1f76c00d6a37bcdfabd986288bff1f873deb38b34 not found: ID does not exist" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.472716 4750 scope.go:117] "RemoveContainer" containerID="4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b" Mar 09 20:12:27 crc kubenswrapper[4750]: E0309 20:12:27.473129 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b\": container with ID starting with 4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b not found: ID does not exist" containerID="4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.473183 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b"} err="failed to get container status \"4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b\": rpc error: code = NotFound desc = could not find container \"4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b\": container with ID starting with 4f9e30cc2ec3afdec36d13ea7a8f23236dad3eee6e662ab4c868fa6c0b72374b not found: ID does not exist" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.473202 4750 scope.go:117] "RemoveContainer" containerID="22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c" Mar 09 20:12:27 crc kubenswrapper[4750]: E0309 20:12:27.473559 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c\": container with ID starting with 22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c not found: ID does not exist" containerID="22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c" Mar 09 20:12:27 crc kubenswrapper[4750]: I0309 20:12:27.473603 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c"} err="failed to get container status \"22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c\": rpc error: code = NotFound desc = could not find container \"22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c\": container with ID starting with 22b4aa887d4b4d394ee573d34d4753677caad98b454a5fd550eccb70bac8089c not found: ID does not exist" Mar 09 20:12:29 crc kubenswrapper[4750]: I0309 20:12:29.390005 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" path="/var/lib/kubelet/pods/7522e5a9-274a-4a16-b508-8418e65bebaf/volumes" Mar 09 20:12:33 crc kubenswrapper[4750]: I0309 20:12:33.374062 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:12:33 crc kubenswrapper[4750]: E0309 20:12:33.375563 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:12:45 crc kubenswrapper[4750]: I0309 20:12:45.374589 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:12:45 crc kubenswrapper[4750]: E0309 20:12:45.375249 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.254278 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mrsst"] Mar 09 20:12:52 crc kubenswrapper[4750]: E0309 20:12:52.255666 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="registry-server" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.255682 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="registry-server" Mar 09 20:12:52 crc kubenswrapper[4750]: E0309 20:12:52.255711 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="extract-utilities" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.255718 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="extract-utilities" Mar 09 20:12:52 crc kubenswrapper[4750]: E0309 20:12:52.255734 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="extract-content" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.255741 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="extract-content" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.255934 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7522e5a9-274a-4a16-b508-8418e65bebaf" containerName="registry-server" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.257317 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.301166 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mrsst"] Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.336239 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-catalog-content\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.336293 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7mcz\" (UniqueName: \"kubernetes.io/projected/2816a158-1d33-4dc8-b78a-8195d23894cc-kube-api-access-x7mcz\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.336438 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-utilities\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.438750 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-catalog-content\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.438797 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7mcz\" (UniqueName: \"kubernetes.io/projected/2816a158-1d33-4dc8-b78a-8195d23894cc-kube-api-access-x7mcz\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.438884 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-utilities\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.439471 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-catalog-content\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.439735 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-utilities\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.467321 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7mcz\" (UniqueName: \"kubernetes.io/projected/2816a158-1d33-4dc8-b78a-8195d23894cc-kube-api-access-x7mcz\") pod \"redhat-operators-mrsst\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.580167 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:12:52 crc kubenswrapper[4750]: I0309 20:12:52.653892 4750 scope.go:117] "RemoveContainer" containerID="7ea1a4f9921c57cac076c9dd63d7e0247057c909fe0a8dd09019c48f1e42c52d" Mar 09 20:12:53 crc kubenswrapper[4750]: I0309 20:12:53.138188 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mrsst"] Mar 09 20:12:53 crc kubenswrapper[4750]: I0309 20:12:53.658713 4750 generic.go:334] "Generic (PLEG): container finished" podID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerID="1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013" exitCode=0 Mar 09 20:12:53 crc kubenswrapper[4750]: I0309 20:12:53.658815 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrsst" event={"ID":"2816a158-1d33-4dc8-b78a-8195d23894cc","Type":"ContainerDied","Data":"1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013"} Mar 09 20:12:53 crc kubenswrapper[4750]: I0309 20:12:53.658965 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrsst" event={"ID":"2816a158-1d33-4dc8-b78a-8195d23894cc","Type":"ContainerStarted","Data":"f96f8f5ffbc545615492e92def941d22fb883f997567acbe088d3b6b63c7d69a"} Mar 09 20:12:53 crc kubenswrapper[4750]: I0309 20:12:53.662182 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 20:12:55 crc kubenswrapper[4750]: I0309 20:12:55.687836 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrsst" event={"ID":"2816a158-1d33-4dc8-b78a-8195d23894cc","Type":"ContainerStarted","Data":"e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587"} Mar 09 20:12:59 crc kubenswrapper[4750]: I0309 20:12:59.400941 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:12:59 crc kubenswrapper[4750]: E0309 20:12:59.401924 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:12:59 crc kubenswrapper[4750]: I0309 20:12:59.736319 4750 generic.go:334] "Generic (PLEG): container finished" podID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerID="e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587" exitCode=0 Mar 09 20:12:59 crc kubenswrapper[4750]: I0309 20:12:59.736402 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrsst" event={"ID":"2816a158-1d33-4dc8-b78a-8195d23894cc","Type":"ContainerDied","Data":"e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587"} Mar 09 20:13:00 crc kubenswrapper[4750]: I0309 20:13:00.748411 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrsst" event={"ID":"2816a158-1d33-4dc8-b78a-8195d23894cc","Type":"ContainerStarted","Data":"4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf"} Mar 09 20:13:00 crc kubenswrapper[4750]: I0309 20:13:00.775850 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mrsst" podStartSLOduration=2.232898056 podStartE2EDuration="8.775822485s" podCreationTimestamp="2026-03-09 20:12:52 +0000 UTC" firstStartedPulling="2026-03-09 20:12:53.661952325 +0000 UTC m=+6455.004424723" lastFinishedPulling="2026-03-09 20:13:00.204876754 +0000 UTC m=+6461.547349152" observedRunningTime="2026-03-09 20:13:00.768233149 +0000 UTC m=+6462.110705597" watchObservedRunningTime="2026-03-09 20:13:00.775822485 +0000 UTC m=+6462.118294913" Mar 09 20:13:02 crc kubenswrapper[4750]: I0309 20:13:02.580661 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:13:02 crc kubenswrapper[4750]: I0309 20:13:02.580988 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:13:03 crc kubenswrapper[4750]: I0309 20:13:03.637578 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mrsst" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="registry-server" probeResult="failure" output=< Mar 09 20:13:03 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 20:13:03 crc kubenswrapper[4750]: > Mar 09 20:13:12 crc kubenswrapper[4750]: I0309 20:13:12.373778 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:13:12 crc kubenswrapper[4750]: E0309 20:13:12.374653 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:13:12 crc kubenswrapper[4750]: I0309 20:13:12.625885 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:13:12 crc kubenswrapper[4750]: I0309 20:13:12.676774 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:13:12 crc kubenswrapper[4750]: I0309 20:13:12.862459 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mrsst"] Mar 09 20:13:13 crc kubenswrapper[4750]: I0309 20:13:13.870299 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mrsst" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="registry-server" containerID="cri-o://4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf" gracePeriod=2 Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.339775 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.507741 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-utilities\") pod \"2816a158-1d33-4dc8-b78a-8195d23894cc\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.507853 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7mcz\" (UniqueName: \"kubernetes.io/projected/2816a158-1d33-4dc8-b78a-8195d23894cc-kube-api-access-x7mcz\") pod \"2816a158-1d33-4dc8-b78a-8195d23894cc\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.508035 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-catalog-content\") pod \"2816a158-1d33-4dc8-b78a-8195d23894cc\" (UID: \"2816a158-1d33-4dc8-b78a-8195d23894cc\") " Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.508613 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-utilities" (OuterVolumeSpecName: "utilities") pod "2816a158-1d33-4dc8-b78a-8195d23894cc" (UID: "2816a158-1d33-4dc8-b78a-8195d23894cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.509803 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.513685 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2816a158-1d33-4dc8-b78a-8195d23894cc-kube-api-access-x7mcz" (OuterVolumeSpecName: "kube-api-access-x7mcz") pod "2816a158-1d33-4dc8-b78a-8195d23894cc" (UID: "2816a158-1d33-4dc8-b78a-8195d23894cc"). InnerVolumeSpecName "kube-api-access-x7mcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.611971 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7mcz\" (UniqueName: \"kubernetes.io/projected/2816a158-1d33-4dc8-b78a-8195d23894cc-kube-api-access-x7mcz\") on node \"crc\" DevicePath \"\"" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.646342 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2816a158-1d33-4dc8-b78a-8195d23894cc" (UID: "2816a158-1d33-4dc8-b78a-8195d23894cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.713753 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2816a158-1d33-4dc8-b78a-8195d23894cc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.881382 4750 generic.go:334] "Generic (PLEG): container finished" podID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerID="4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf" exitCode=0 Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.881424 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrsst" event={"ID":"2816a158-1d33-4dc8-b78a-8195d23894cc","Type":"ContainerDied","Data":"4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf"} Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.881447 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrsst" event={"ID":"2816a158-1d33-4dc8-b78a-8195d23894cc","Type":"ContainerDied","Data":"f96f8f5ffbc545615492e92def941d22fb883f997567acbe088d3b6b63c7d69a"} Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.881464 4750 scope.go:117] "RemoveContainer" containerID="4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.881466 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrsst" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.917517 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mrsst"] Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.931218 4750 scope.go:117] "RemoveContainer" containerID="e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587" Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.934080 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mrsst"] Mar 09 20:13:14 crc kubenswrapper[4750]: I0309 20:13:14.964838 4750 scope.go:117] "RemoveContainer" containerID="1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013" Mar 09 20:13:15 crc kubenswrapper[4750]: I0309 20:13:15.023208 4750 scope.go:117] "RemoveContainer" containerID="4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf" Mar 09 20:13:15 crc kubenswrapper[4750]: E0309 20:13:15.023599 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf\": container with ID starting with 4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf not found: ID does not exist" containerID="4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf" Mar 09 20:13:15 crc kubenswrapper[4750]: I0309 20:13:15.023649 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf"} err="failed to get container status \"4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf\": rpc error: code = NotFound desc = could not find container \"4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf\": container with ID starting with 4da61e66661bb8e7ec99fa4c1c3be080fb89ccaf696d481a671fabbcf63380bf not found: ID does not exist" Mar 09 20:13:15 crc kubenswrapper[4750]: I0309 20:13:15.023672 4750 scope.go:117] "RemoveContainer" containerID="e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587" Mar 09 20:13:15 crc kubenswrapper[4750]: E0309 20:13:15.024010 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587\": container with ID starting with e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587 not found: ID does not exist" containerID="e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587" Mar 09 20:13:15 crc kubenswrapper[4750]: I0309 20:13:15.024049 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587"} err="failed to get container status \"e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587\": rpc error: code = NotFound desc = could not find container \"e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587\": container with ID starting with e1af6d84215aa2a57f3ff80f25584df332704f3520ff6a083747dfa843ca2587 not found: ID does not exist" Mar 09 20:13:15 crc kubenswrapper[4750]: I0309 20:13:15.024077 4750 scope.go:117] "RemoveContainer" containerID="1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013" Mar 09 20:13:15 crc kubenswrapper[4750]: E0309 20:13:15.024378 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013\": container with ID starting with 1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013 not found: ID does not exist" containerID="1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013" Mar 09 20:13:15 crc kubenswrapper[4750]: I0309 20:13:15.024398 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013"} err="failed to get container status \"1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013\": rpc error: code = NotFound desc = could not find container \"1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013\": container with ID starting with 1d886001eb0835410d02f2382d79655c26490b9af6a4f2de1f947ece90afe013 not found: ID does not exist" Mar 09 20:13:15 crc kubenswrapper[4750]: I0309 20:13:15.393152 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" path="/var/lib/kubelet/pods/2816a158-1d33-4dc8-b78a-8195d23894cc/volumes" Mar 09 20:13:26 crc kubenswrapper[4750]: I0309 20:13:26.372656 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:13:26 crc kubenswrapper[4750]: E0309 20:13:26.373256 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:13:37 crc kubenswrapper[4750]: I0309 20:13:37.373956 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:13:37 crc kubenswrapper[4750]: E0309 20:13:37.374577 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:13:50 crc kubenswrapper[4750]: I0309 20:13:50.373606 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:13:50 crc kubenswrapper[4750]: E0309 20:13:50.374449 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:13:51 crc kubenswrapper[4750]: I0309 20:13:51.743537 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:13:51 crc kubenswrapper[4750]: I0309 20:13:51.743892 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.159932 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551454-7vvml"] Mar 09 20:14:00 crc kubenswrapper[4750]: E0309 20:14:00.161331 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="extract-content" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.161347 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="extract-content" Mar 09 20:14:00 crc kubenswrapper[4750]: E0309 20:14:00.161370 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="extract-utilities" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.161379 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="extract-utilities" Mar 09 20:14:00 crc kubenswrapper[4750]: E0309 20:14:00.161401 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="registry-server" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.161407 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="registry-server" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.162458 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2816a158-1d33-4dc8-b78a-8195d23894cc" containerName="registry-server" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.163972 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551454-7vvml" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.166017 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.166213 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.167992 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.172859 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551454-7vvml"] Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.202786 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bc9l\" (UniqueName: \"kubernetes.io/projected/cdf60d44-00ad-45fb-b258-c309af142f3b-kube-api-access-5bc9l\") pod \"auto-csr-approver-29551454-7vvml\" (UID: \"cdf60d44-00ad-45fb-b258-c309af142f3b\") " pod="openshift-infra/auto-csr-approver-29551454-7vvml" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.304573 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bc9l\" (UniqueName: \"kubernetes.io/projected/cdf60d44-00ad-45fb-b258-c309af142f3b-kube-api-access-5bc9l\") pod \"auto-csr-approver-29551454-7vvml\" (UID: \"cdf60d44-00ad-45fb-b258-c309af142f3b\") " pod="openshift-infra/auto-csr-approver-29551454-7vvml" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.325215 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bc9l\" (UniqueName: \"kubernetes.io/projected/cdf60d44-00ad-45fb-b258-c309af142f3b-kube-api-access-5bc9l\") pod \"auto-csr-approver-29551454-7vvml\" (UID: \"cdf60d44-00ad-45fb-b258-c309af142f3b\") " pod="openshift-infra/auto-csr-approver-29551454-7vvml" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.499181 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551454-7vvml" Mar 09 20:14:00 crc kubenswrapper[4750]: I0309 20:14:00.965038 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551454-7vvml"] Mar 09 20:14:01 crc kubenswrapper[4750]: I0309 20:14:01.370521 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551454-7vvml" event={"ID":"cdf60d44-00ad-45fb-b258-c309af142f3b","Type":"ContainerStarted","Data":"541b8da635dbe7cf4d3441ea08933e81dbaf06351105442803422bdb72b36a91"} Mar 09 20:14:03 crc kubenswrapper[4750]: I0309 20:14:03.391812 4750 generic.go:334] "Generic (PLEG): container finished" podID="cdf60d44-00ad-45fb-b258-c309af142f3b" containerID="fa36304e4581e5b87c02d4446a9a18331d63acaee77d8fdbc266ebd9112f4ee5" exitCode=0 Mar 09 20:14:03 crc kubenswrapper[4750]: I0309 20:14:03.392127 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551454-7vvml" event={"ID":"cdf60d44-00ad-45fb-b258-c309af142f3b","Type":"ContainerDied","Data":"fa36304e4581e5b87c02d4446a9a18331d63acaee77d8fdbc266ebd9112f4ee5"} Mar 09 20:14:04 crc kubenswrapper[4750]: I0309 20:14:04.373479 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:14:04 crc kubenswrapper[4750]: E0309 20:14:04.374268 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:14:04 crc kubenswrapper[4750]: I0309 20:14:04.764995 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551454-7vvml" Mar 09 20:14:04 crc kubenswrapper[4750]: I0309 20:14:04.903032 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bc9l\" (UniqueName: \"kubernetes.io/projected/cdf60d44-00ad-45fb-b258-c309af142f3b-kube-api-access-5bc9l\") pod \"cdf60d44-00ad-45fb-b258-c309af142f3b\" (UID: \"cdf60d44-00ad-45fb-b258-c309af142f3b\") " Mar 09 20:14:04 crc kubenswrapper[4750]: I0309 20:14:04.909522 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf60d44-00ad-45fb-b258-c309af142f3b-kube-api-access-5bc9l" (OuterVolumeSpecName: "kube-api-access-5bc9l") pod "cdf60d44-00ad-45fb-b258-c309af142f3b" (UID: "cdf60d44-00ad-45fb-b258-c309af142f3b"). InnerVolumeSpecName "kube-api-access-5bc9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:14:05 crc kubenswrapper[4750]: I0309 20:14:05.005955 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bc9l\" (UniqueName: \"kubernetes.io/projected/cdf60d44-00ad-45fb-b258-c309af142f3b-kube-api-access-5bc9l\") on node \"crc\" DevicePath \"\"" Mar 09 20:14:05 crc kubenswrapper[4750]: I0309 20:14:05.422687 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551454-7vvml" event={"ID":"cdf60d44-00ad-45fb-b258-c309af142f3b","Type":"ContainerDied","Data":"541b8da635dbe7cf4d3441ea08933e81dbaf06351105442803422bdb72b36a91"} Mar 09 20:14:05 crc kubenswrapper[4750]: I0309 20:14:05.422741 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="541b8da635dbe7cf4d3441ea08933e81dbaf06351105442803422bdb72b36a91" Mar 09 20:14:05 crc kubenswrapper[4750]: I0309 20:14:05.422782 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551454-7vvml" Mar 09 20:14:05 crc kubenswrapper[4750]: I0309 20:14:05.830604 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551448-ds2nd"] Mar 09 20:14:05 crc kubenswrapper[4750]: I0309 20:14:05.839527 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551448-ds2nd"] Mar 09 20:14:07 crc kubenswrapper[4750]: I0309 20:14:07.385380 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee4da78a-53ca-4ac8-b128-b943d6e58439" path="/var/lib/kubelet/pods/ee4da78a-53ca-4ac8-b128-b943d6e58439/volumes" Mar 09 20:14:19 crc kubenswrapper[4750]: I0309 20:14:19.373887 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:14:19 crc kubenswrapper[4750]: E0309 20:14:19.374706 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:14:21 crc kubenswrapper[4750]: I0309 20:14:21.744399 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:14:21 crc kubenswrapper[4750]: I0309 20:14:21.744943 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:14:34 crc kubenswrapper[4750]: I0309 20:14:34.374352 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:14:34 crc kubenswrapper[4750]: E0309 20:14:34.375439 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:14:49 crc kubenswrapper[4750]: I0309 20:14:49.385502 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:14:49 crc kubenswrapper[4750]: E0309 20:14:49.387209 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.743689 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.744078 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.744137 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.745065 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1aee94e71f7a10634c557910d5c09ce69ff26474679f02e5aab3816f75d7901e"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.745134 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://1aee94e71f7a10634c557910d5c09ce69ff26474679f02e5aab3816f75d7901e" gracePeriod=600 Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.947288 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="1aee94e71f7a10634c557910d5c09ce69ff26474679f02e5aab3816f75d7901e" exitCode=0 Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.947340 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"1aee94e71f7a10634c557910d5c09ce69ff26474679f02e5aab3816f75d7901e"} Mar 09 20:14:51 crc kubenswrapper[4750]: I0309 20:14:51.947377 4750 scope.go:117] "RemoveContainer" containerID="a6fb9ed58b14d0622197f2de69109645c7262033b15e09fe3b3cc2ecab1d327d" Mar 09 20:14:52 crc kubenswrapper[4750]: I0309 20:14:52.860320 4750 scope.go:117] "RemoveContainer" containerID="e2caa00cfb6c067920b02dc8fda6cac663c433b844a2de3f5776f2db93503ec7" Mar 09 20:14:52 crc kubenswrapper[4750]: I0309 20:14:52.972392 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8"} Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.538908 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pkxq8/must-gather-wnhhg"] Mar 09 20:14:53 crc kubenswrapper[4750]: E0309 20:14:53.539534 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf60d44-00ad-45fb-b258-c309af142f3b" containerName="oc" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.539550 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf60d44-00ad-45fb-b258-c309af142f3b" containerName="oc" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.539793 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf60d44-00ad-45fb-b258-c309af142f3b" containerName="oc" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.540789 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.542978 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pkxq8"/"kube-root-ca.crt" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.543187 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pkxq8"/"default-dockercfg-7hkqv" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.548924 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pkxq8"/"openshift-service-ca.crt" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.553324 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pkxq8/must-gather-wnhhg"] Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.675091 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/07185cb9-e46d-4abc-9ad2-55b113c0a29f-must-gather-output\") pod \"must-gather-wnhhg\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.675186 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xvnx\" (UniqueName: \"kubernetes.io/projected/07185cb9-e46d-4abc-9ad2-55b113c0a29f-kube-api-access-9xvnx\") pod \"must-gather-wnhhg\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.777832 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/07185cb9-e46d-4abc-9ad2-55b113c0a29f-must-gather-output\") pod \"must-gather-wnhhg\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.777911 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xvnx\" (UniqueName: \"kubernetes.io/projected/07185cb9-e46d-4abc-9ad2-55b113c0a29f-kube-api-access-9xvnx\") pod \"must-gather-wnhhg\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.778293 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/07185cb9-e46d-4abc-9ad2-55b113c0a29f-must-gather-output\") pod \"must-gather-wnhhg\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.806404 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xvnx\" (UniqueName: \"kubernetes.io/projected/07185cb9-e46d-4abc-9ad2-55b113c0a29f-kube-api-access-9xvnx\") pod \"must-gather-wnhhg\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:53 crc kubenswrapper[4750]: I0309 20:14:53.858284 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:14:54 crc kubenswrapper[4750]: I0309 20:14:54.359402 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pkxq8/must-gather-wnhhg"] Mar 09 20:14:54 crc kubenswrapper[4750]: W0309 20:14:54.369853 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07185cb9_e46d_4abc_9ad2_55b113c0a29f.slice/crio-01338f75778f8f154ca816b93146f57a0b41f0fead6a4d4e9592b392f0544b1b WatchSource:0}: Error finding container 01338f75778f8f154ca816b93146f57a0b41f0fead6a4d4e9592b392f0544b1b: Status 404 returned error can't find the container with id 01338f75778f8f154ca816b93146f57a0b41f0fead6a4d4e9592b392f0544b1b Mar 09 20:14:55 crc kubenswrapper[4750]: I0309 20:14:55.010057 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" event={"ID":"07185cb9-e46d-4abc-9ad2-55b113c0a29f","Type":"ContainerStarted","Data":"01338f75778f8f154ca816b93146f57a0b41f0fead6a4d4e9592b392f0544b1b"} Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.147512 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg"] Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.149620 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.153937 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.154031 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.161360 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg"] Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.330557 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-config-volume\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.330733 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx6dj\" (UniqueName: \"kubernetes.io/projected/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-kube-api-access-jx6dj\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.330889 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-secret-volume\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.373919 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:15:00 crc kubenswrapper[4750]: E0309 20:15:00.374377 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.432351 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-secret-volume\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.432430 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-config-volume\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.432475 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx6dj\" (UniqueName: \"kubernetes.io/projected/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-kube-api-access-jx6dj\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.433866 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-config-volume\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.443368 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-secret-volume\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.452695 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx6dj\" (UniqueName: \"kubernetes.io/projected/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-kube-api-access-jx6dj\") pod \"collect-profiles-29551455-qbpfg\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:00 crc kubenswrapper[4750]: I0309 20:15:00.482768 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:01 crc kubenswrapper[4750]: I0309 20:15:01.067006 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" event={"ID":"07185cb9-e46d-4abc-9ad2-55b113c0a29f","Type":"ContainerStarted","Data":"8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99"} Mar 09 20:15:01 crc kubenswrapper[4750]: I0309 20:15:01.081797 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg"] Mar 09 20:15:01 crc kubenswrapper[4750]: W0309 20:15:01.084324 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedff4d43_6b61_41a9_a69d_95c2eb5de5ce.slice/crio-b404382c4b26993046af86e01f6468fd9db877af71d10b62db635876974c4e2f WatchSource:0}: Error finding container b404382c4b26993046af86e01f6468fd9db877af71d10b62db635876974c4e2f: Status 404 returned error can't find the container with id b404382c4b26993046af86e01f6468fd9db877af71d10b62db635876974c4e2f Mar 09 20:15:02 crc kubenswrapper[4750]: I0309 20:15:02.078046 4750 generic.go:334] "Generic (PLEG): container finished" podID="edff4d43-6b61-41a9-a69d-95c2eb5de5ce" containerID="f0dd162993f61f25354de4b5cf24d9174587f74a08802ba2d8d648ee9326e289" exitCode=0 Mar 09 20:15:02 crc kubenswrapper[4750]: I0309 20:15:02.078108 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" event={"ID":"edff4d43-6b61-41a9-a69d-95c2eb5de5ce","Type":"ContainerDied","Data":"f0dd162993f61f25354de4b5cf24d9174587f74a08802ba2d8d648ee9326e289"} Mar 09 20:15:02 crc kubenswrapper[4750]: I0309 20:15:02.078764 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" event={"ID":"edff4d43-6b61-41a9-a69d-95c2eb5de5ce","Type":"ContainerStarted","Data":"b404382c4b26993046af86e01f6468fd9db877af71d10b62db635876974c4e2f"} Mar 09 20:15:02 crc kubenswrapper[4750]: I0309 20:15:02.080606 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" event={"ID":"07185cb9-e46d-4abc-9ad2-55b113c0a29f","Type":"ContainerStarted","Data":"c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb"} Mar 09 20:15:02 crc kubenswrapper[4750]: I0309 20:15:02.128989 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" podStartSLOduration=2.693907442 podStartE2EDuration="9.128970043s" podCreationTimestamp="2026-03-09 20:14:53 +0000 UTC" firstStartedPulling="2026-03-09 20:14:54.372239882 +0000 UTC m=+6575.714712280" lastFinishedPulling="2026-03-09 20:15:00.807302483 +0000 UTC m=+6582.149774881" observedRunningTime="2026-03-09 20:15:02.121406608 +0000 UTC m=+6583.463879046" watchObservedRunningTime="2026-03-09 20:15:02.128970043 +0000 UTC m=+6583.471442451" Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.431376 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.517663 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-secret-volume\") pod \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.517826 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx6dj\" (UniqueName: \"kubernetes.io/projected/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-kube-api-access-jx6dj\") pod \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.517875 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-config-volume\") pod \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\" (UID: \"edff4d43-6b61-41a9-a69d-95c2eb5de5ce\") " Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.518726 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-config-volume" (OuterVolumeSpecName: "config-volume") pod "edff4d43-6b61-41a9-a69d-95c2eb5de5ce" (UID: "edff4d43-6b61-41a9-a69d-95c2eb5de5ce"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.536000 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "edff4d43-6b61-41a9-a69d-95c2eb5de5ce" (UID: "edff4d43-6b61-41a9-a69d-95c2eb5de5ce"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.536051 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-kube-api-access-jx6dj" (OuterVolumeSpecName: "kube-api-access-jx6dj") pod "edff4d43-6b61-41a9-a69d-95c2eb5de5ce" (UID: "edff4d43-6b61-41a9-a69d-95c2eb5de5ce"). InnerVolumeSpecName "kube-api-access-jx6dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.620120 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.620155 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx6dj\" (UniqueName: \"kubernetes.io/projected/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-kube-api-access-jx6dj\") on node \"crc\" DevicePath \"\"" Mar 09 20:15:03 crc kubenswrapper[4750]: I0309 20:15:03.620165 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edff4d43-6b61-41a9-a69d-95c2eb5de5ce-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 20:15:04 crc kubenswrapper[4750]: I0309 20:15:04.100987 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" event={"ID":"edff4d43-6b61-41a9-a69d-95c2eb5de5ce","Type":"ContainerDied","Data":"b404382c4b26993046af86e01f6468fd9db877af71d10b62db635876974c4e2f"} Mar 09 20:15:04 crc kubenswrapper[4750]: I0309 20:15:04.101013 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29551455-qbpfg" Mar 09 20:15:04 crc kubenswrapper[4750]: I0309 20:15:04.101020 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b404382c4b26993046af86e01f6468fd9db877af71d10b62db635876974c4e2f" Mar 09 20:15:04 crc kubenswrapper[4750]: I0309 20:15:04.509126 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq"] Mar 09 20:15:04 crc kubenswrapper[4750]: I0309 20:15:04.518018 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29551410-s8rcq"] Mar 09 20:15:05 crc kubenswrapper[4750]: I0309 20:15:05.389410 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4" path="/var/lib/kubelet/pods/2b98e5a7-bfd5-40e7-b35a-4b6e2649bcf4/volumes" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.275377 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pkxq8/crc-debug-q96kc"] Mar 09 20:15:06 crc kubenswrapper[4750]: E0309 20:15:06.276339 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edff4d43-6b61-41a9-a69d-95c2eb5de5ce" containerName="collect-profiles" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.276358 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="edff4d43-6b61-41a9-a69d-95c2eb5de5ce" containerName="collect-profiles" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.276596 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="edff4d43-6b61-41a9-a69d-95c2eb5de5ce" containerName="collect-profiles" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.277435 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.372994 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02087f90-1033-43d5-bc20-5b549c024aab-host\") pod \"crc-debug-q96kc\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.373405 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tsg4\" (UniqueName: \"kubernetes.io/projected/02087f90-1033-43d5-bc20-5b549c024aab-kube-api-access-7tsg4\") pod \"crc-debug-q96kc\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.475136 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tsg4\" (UniqueName: \"kubernetes.io/projected/02087f90-1033-43d5-bc20-5b549c024aab-kube-api-access-7tsg4\") pod \"crc-debug-q96kc\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.475399 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02087f90-1033-43d5-bc20-5b549c024aab-host\") pod \"crc-debug-q96kc\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.475527 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02087f90-1033-43d5-bc20-5b549c024aab-host\") pod \"crc-debug-q96kc\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.504814 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tsg4\" (UniqueName: \"kubernetes.io/projected/02087f90-1033-43d5-bc20-5b549c024aab-kube-api-access-7tsg4\") pod \"crc-debug-q96kc\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: I0309 20:15:06.599982 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:06 crc kubenswrapper[4750]: W0309 20:15:06.656759 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02087f90_1033_43d5_bc20_5b549c024aab.slice/crio-cbd14109ece0d0fd1d56983f3e1560fbc100ba3d9c5d85dd3cffb1013f549c07 WatchSource:0}: Error finding container cbd14109ece0d0fd1d56983f3e1560fbc100ba3d9c5d85dd3cffb1013f549c07: Status 404 returned error can't find the container with id cbd14109ece0d0fd1d56983f3e1560fbc100ba3d9c5d85dd3cffb1013f549c07 Mar 09 20:15:07 crc kubenswrapper[4750]: I0309 20:15:07.129922 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" event={"ID":"02087f90-1033-43d5-bc20-5b549c024aab","Type":"ContainerStarted","Data":"cbd14109ece0d0fd1d56983f3e1560fbc100ba3d9c5d85dd3cffb1013f549c07"} Mar 09 20:15:15 crc kubenswrapper[4750]: I0309 20:15:15.374043 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:15:15 crc kubenswrapper[4750]: E0309 20:15:15.374733 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:15:19 crc kubenswrapper[4750]: I0309 20:15:19.251412 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" event={"ID":"02087f90-1033-43d5-bc20-5b549c024aab","Type":"ContainerStarted","Data":"8cb417ca3b4483af6d5e4d6278de0f9afd375dfe0eb3bd6c1a5705b728104a7d"} Mar 09 20:15:19 crc kubenswrapper[4750]: I0309 20:15:19.271532 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" podStartSLOduration=1.295853938 podStartE2EDuration="13.271516817s" podCreationTimestamp="2026-03-09 20:15:06 +0000 UTC" firstStartedPulling="2026-03-09 20:15:06.660353647 +0000 UTC m=+6588.002826045" lastFinishedPulling="2026-03-09 20:15:18.636016526 +0000 UTC m=+6599.978488924" observedRunningTime="2026-03-09 20:15:19.267393976 +0000 UTC m=+6600.609866394" watchObservedRunningTime="2026-03-09 20:15:19.271516817 +0000 UTC m=+6600.613989215" Mar 09 20:15:27 crc kubenswrapper[4750]: I0309 20:15:27.373506 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:15:27 crc kubenswrapper[4750]: E0309 20:15:27.374295 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:15:33 crc kubenswrapper[4750]: I0309 20:15:33.391926 4750 generic.go:334] "Generic (PLEG): container finished" podID="02087f90-1033-43d5-bc20-5b549c024aab" containerID="8cb417ca3b4483af6d5e4d6278de0f9afd375dfe0eb3bd6c1a5705b728104a7d" exitCode=0 Mar 09 20:15:33 crc kubenswrapper[4750]: I0309 20:15:33.392014 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" event={"ID":"02087f90-1033-43d5-bc20-5b549c024aab","Type":"ContainerDied","Data":"8cb417ca3b4483af6d5e4d6278de0f9afd375dfe0eb3bd6c1a5705b728104a7d"} Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.523405 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.556667 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pkxq8/crc-debug-q96kc"] Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.565213 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pkxq8/crc-debug-q96kc"] Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.633029 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02087f90-1033-43d5-bc20-5b549c024aab-host\") pod \"02087f90-1033-43d5-bc20-5b549c024aab\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.633223 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tsg4\" (UniqueName: \"kubernetes.io/projected/02087f90-1033-43d5-bc20-5b549c024aab-kube-api-access-7tsg4\") pod \"02087f90-1033-43d5-bc20-5b549c024aab\" (UID: \"02087f90-1033-43d5-bc20-5b549c024aab\") " Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.633393 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02087f90-1033-43d5-bc20-5b549c024aab-host" (OuterVolumeSpecName: "host") pod "02087f90-1033-43d5-bc20-5b549c024aab" (UID: "02087f90-1033-43d5-bc20-5b549c024aab"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.633847 4750 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02087f90-1033-43d5-bc20-5b549c024aab-host\") on node \"crc\" DevicePath \"\"" Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.639743 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02087f90-1033-43d5-bc20-5b549c024aab-kube-api-access-7tsg4" (OuterVolumeSpecName: "kube-api-access-7tsg4") pod "02087f90-1033-43d5-bc20-5b549c024aab" (UID: "02087f90-1033-43d5-bc20-5b549c024aab"). InnerVolumeSpecName "kube-api-access-7tsg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:15:34 crc kubenswrapper[4750]: I0309 20:15:34.736295 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tsg4\" (UniqueName: \"kubernetes.io/projected/02087f90-1033-43d5-bc20-5b549c024aab-kube-api-access-7tsg4\") on node \"crc\" DevicePath \"\"" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.386512 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02087f90-1033-43d5-bc20-5b549c024aab" path="/var/lib/kubelet/pods/02087f90-1033-43d5-bc20-5b549c024aab/volumes" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.416161 4750 scope.go:117] "RemoveContainer" containerID="8cb417ca3b4483af6d5e4d6278de0f9afd375dfe0eb3bd6c1a5705b728104a7d" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.416254 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-q96kc" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.809009 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pkxq8/crc-debug-g2xdh"] Mar 09 20:15:35 crc kubenswrapper[4750]: E0309 20:15:35.810006 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02087f90-1033-43d5-bc20-5b549c024aab" containerName="container-00" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.810075 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="02087f90-1033-43d5-bc20-5b549c024aab" containerName="container-00" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.810336 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="02087f90-1033-43d5-bc20-5b549c024aab" containerName="container-00" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.811030 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.861073 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6543c9b-8445-462b-bff4-4d834830dd80-host\") pod \"crc-debug-g2xdh\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.861198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf8qp\" (UniqueName: \"kubernetes.io/projected/b6543c9b-8445-462b-bff4-4d834830dd80-kube-api-access-xf8qp\") pod \"crc-debug-g2xdh\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.962544 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6543c9b-8445-462b-bff4-4d834830dd80-host\") pod \"crc-debug-g2xdh\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.962650 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6543c9b-8445-462b-bff4-4d834830dd80-host\") pod \"crc-debug-g2xdh\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.962676 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf8qp\" (UniqueName: \"kubernetes.io/projected/b6543c9b-8445-462b-bff4-4d834830dd80-kube-api-access-xf8qp\") pod \"crc-debug-g2xdh\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:35 crc kubenswrapper[4750]: I0309 20:15:35.980738 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf8qp\" (UniqueName: \"kubernetes.io/projected/b6543c9b-8445-462b-bff4-4d834830dd80-kube-api-access-xf8qp\") pod \"crc-debug-g2xdh\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:36 crc kubenswrapper[4750]: I0309 20:15:36.126612 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:36 crc kubenswrapper[4750]: I0309 20:15:36.429260 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" event={"ID":"b6543c9b-8445-462b-bff4-4d834830dd80","Type":"ContainerStarted","Data":"f23ebd7daf693e38f1fa42c8b45116df50dfd9611e2b8bd4ac6db2b4ed64daa9"} Mar 09 20:15:37 crc kubenswrapper[4750]: I0309 20:15:37.439888 4750 generic.go:334] "Generic (PLEG): container finished" podID="b6543c9b-8445-462b-bff4-4d834830dd80" containerID="16880d93c49df0cc2d95688e8eb63c92efaa1a877418b59e847fe0844bf29221" exitCode=1 Mar 09 20:15:37 crc kubenswrapper[4750]: I0309 20:15:37.439990 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" event={"ID":"b6543c9b-8445-462b-bff4-4d834830dd80","Type":"ContainerDied","Data":"16880d93c49df0cc2d95688e8eb63c92efaa1a877418b59e847fe0844bf29221"} Mar 09 20:15:37 crc kubenswrapper[4750]: I0309 20:15:37.481565 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pkxq8/crc-debug-g2xdh"] Mar 09 20:15:37 crc kubenswrapper[4750]: I0309 20:15:37.495131 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pkxq8/crc-debug-g2xdh"] Mar 09 20:15:38 crc kubenswrapper[4750]: I0309 20:15:38.575012 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:38 crc kubenswrapper[4750]: I0309 20:15:38.612363 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf8qp\" (UniqueName: \"kubernetes.io/projected/b6543c9b-8445-462b-bff4-4d834830dd80-kube-api-access-xf8qp\") pod \"b6543c9b-8445-462b-bff4-4d834830dd80\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " Mar 09 20:15:38 crc kubenswrapper[4750]: I0309 20:15:38.612426 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6543c9b-8445-462b-bff4-4d834830dd80-host\") pod \"b6543c9b-8445-462b-bff4-4d834830dd80\" (UID: \"b6543c9b-8445-462b-bff4-4d834830dd80\") " Mar 09 20:15:38 crc kubenswrapper[4750]: I0309 20:15:38.612591 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6543c9b-8445-462b-bff4-4d834830dd80-host" (OuterVolumeSpecName: "host") pod "b6543c9b-8445-462b-bff4-4d834830dd80" (UID: "b6543c9b-8445-462b-bff4-4d834830dd80"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 20:15:38 crc kubenswrapper[4750]: I0309 20:15:38.613135 4750 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6543c9b-8445-462b-bff4-4d834830dd80-host\") on node \"crc\" DevicePath \"\"" Mar 09 20:15:38 crc kubenswrapper[4750]: I0309 20:15:38.653966 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6543c9b-8445-462b-bff4-4d834830dd80-kube-api-access-xf8qp" (OuterVolumeSpecName: "kube-api-access-xf8qp") pod "b6543c9b-8445-462b-bff4-4d834830dd80" (UID: "b6543c9b-8445-462b-bff4-4d834830dd80"). InnerVolumeSpecName "kube-api-access-xf8qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:15:38 crc kubenswrapper[4750]: I0309 20:15:38.715663 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf8qp\" (UniqueName: \"kubernetes.io/projected/b6543c9b-8445-462b-bff4-4d834830dd80-kube-api-access-xf8qp\") on node \"crc\" DevicePath \"\"" Mar 09 20:15:39 crc kubenswrapper[4750]: I0309 20:15:39.387603 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6543c9b-8445-462b-bff4-4d834830dd80" path="/var/lib/kubelet/pods/b6543c9b-8445-462b-bff4-4d834830dd80/volumes" Mar 09 20:15:39 crc kubenswrapper[4750]: I0309 20:15:39.474367 4750 scope.go:117] "RemoveContainer" containerID="16880d93c49df0cc2d95688e8eb63c92efaa1a877418b59e847fe0844bf29221" Mar 09 20:15:39 crc kubenswrapper[4750]: I0309 20:15:39.474388 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/crc-debug-g2xdh" Mar 09 20:15:42 crc kubenswrapper[4750]: I0309 20:15:42.374520 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:15:42 crc kubenswrapper[4750]: E0309 20:15:42.375400 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:15:52 crc kubenswrapper[4750]: I0309 20:15:52.993372 4750 scope.go:117] "RemoveContainer" containerID="7fd159ad10b8a8cba3952ad573df40441b97914c0b7368c37d12150131783ec6" Mar 09 20:15:57 crc kubenswrapper[4750]: I0309 20:15:57.373456 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:15:57 crc kubenswrapper[4750]: E0309 20:15:57.374334 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.138642 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551456-npxpx"] Mar 09 20:16:00 crc kubenswrapper[4750]: E0309 20:16:00.139572 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6543c9b-8445-462b-bff4-4d834830dd80" containerName="container-00" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.139612 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6543c9b-8445-462b-bff4-4d834830dd80" containerName="container-00" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.139868 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6543c9b-8445-462b-bff4-4d834830dd80" containerName="container-00" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.140598 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551456-npxpx" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.142342 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.142360 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.142858 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.158006 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l5pw\" (UniqueName: \"kubernetes.io/projected/756bf273-1cee-409f-9c91-b45473160359-kube-api-access-9l5pw\") pod \"auto-csr-approver-29551456-npxpx\" (UID: \"756bf273-1cee-409f-9c91-b45473160359\") " pod="openshift-infra/auto-csr-approver-29551456-npxpx" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.160523 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551456-npxpx"] Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.260032 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l5pw\" (UniqueName: \"kubernetes.io/projected/756bf273-1cee-409f-9c91-b45473160359-kube-api-access-9l5pw\") pod \"auto-csr-approver-29551456-npxpx\" (UID: \"756bf273-1cee-409f-9c91-b45473160359\") " pod="openshift-infra/auto-csr-approver-29551456-npxpx" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.278879 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l5pw\" (UniqueName: \"kubernetes.io/projected/756bf273-1cee-409f-9c91-b45473160359-kube-api-access-9l5pw\") pod \"auto-csr-approver-29551456-npxpx\" (UID: \"756bf273-1cee-409f-9c91-b45473160359\") " pod="openshift-infra/auto-csr-approver-29551456-npxpx" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.461494 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551456-npxpx" Mar 09 20:16:00 crc kubenswrapper[4750]: I0309 20:16:00.934617 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551456-npxpx"] Mar 09 20:16:01 crc kubenswrapper[4750]: I0309 20:16:01.685703 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551456-npxpx" event={"ID":"756bf273-1cee-409f-9c91-b45473160359","Type":"ContainerStarted","Data":"3ebf41e2eda5e799fc8461fff861cbfd848502ec98b3d232b332567717be5976"} Mar 09 20:16:02 crc kubenswrapper[4750]: I0309 20:16:02.719799 4750 generic.go:334] "Generic (PLEG): container finished" podID="756bf273-1cee-409f-9c91-b45473160359" containerID="dc23350786883987400b43a67df16f43ed991845ec0e61903a312c09dd49c83e" exitCode=0 Mar 09 20:16:02 crc kubenswrapper[4750]: I0309 20:16:02.720091 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551456-npxpx" event={"ID":"756bf273-1cee-409f-9c91-b45473160359","Type":"ContainerDied","Data":"dc23350786883987400b43a67df16f43ed991845ec0e61903a312c09dd49c83e"} Mar 09 20:16:04 crc kubenswrapper[4750]: I0309 20:16:04.096607 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551456-npxpx" Mar 09 20:16:04 crc kubenswrapper[4750]: I0309 20:16:04.162315 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l5pw\" (UniqueName: \"kubernetes.io/projected/756bf273-1cee-409f-9c91-b45473160359-kube-api-access-9l5pw\") pod \"756bf273-1cee-409f-9c91-b45473160359\" (UID: \"756bf273-1cee-409f-9c91-b45473160359\") " Mar 09 20:16:04 crc kubenswrapper[4750]: I0309 20:16:04.169653 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756bf273-1cee-409f-9c91-b45473160359-kube-api-access-9l5pw" (OuterVolumeSpecName: "kube-api-access-9l5pw") pod "756bf273-1cee-409f-9c91-b45473160359" (UID: "756bf273-1cee-409f-9c91-b45473160359"). InnerVolumeSpecName "kube-api-access-9l5pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:16:04 crc kubenswrapper[4750]: I0309 20:16:04.264373 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l5pw\" (UniqueName: \"kubernetes.io/projected/756bf273-1cee-409f-9c91-b45473160359-kube-api-access-9l5pw\") on node \"crc\" DevicePath \"\"" Mar 09 20:16:04 crc kubenswrapper[4750]: I0309 20:16:04.745478 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551456-npxpx" event={"ID":"756bf273-1cee-409f-9c91-b45473160359","Type":"ContainerDied","Data":"3ebf41e2eda5e799fc8461fff861cbfd848502ec98b3d232b332567717be5976"} Mar 09 20:16:04 crc kubenswrapper[4750]: I0309 20:16:04.745529 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ebf41e2eda5e799fc8461fff861cbfd848502ec98b3d232b332567717be5976" Mar 09 20:16:04 crc kubenswrapper[4750]: I0309 20:16:04.745545 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551456-npxpx" Mar 09 20:16:05 crc kubenswrapper[4750]: I0309 20:16:05.163604 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551450-zn4xs"] Mar 09 20:16:05 crc kubenswrapper[4750]: I0309 20:16:05.173019 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551450-zn4xs"] Mar 09 20:16:05 crc kubenswrapper[4750]: I0309 20:16:05.386414 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c695cb-27f6-4a1d-9065-798fff27f528" path="/var/lib/kubelet/pods/85c695cb-27f6-4a1d-9065-798fff27f528/volumes" Mar 09 20:16:11 crc kubenswrapper[4750]: I0309 20:16:11.374164 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:16:11 crc kubenswrapper[4750]: E0309 20:16:11.375149 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:16:24 crc kubenswrapper[4750]: I0309 20:16:24.375261 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:16:24 crc kubenswrapper[4750]: E0309 20:16:24.376195 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:16:37 crc kubenswrapper[4750]: I0309 20:16:37.374836 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:16:38 crc kubenswrapper[4750]: I0309 20:16:38.079971 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400"} Mar 09 20:16:39 crc kubenswrapper[4750]: I0309 20:16:39.451188 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-79b78f4bf4-klzwx_6116d3b9-23b6-44a7-bf39-3b139a2a771e/barbican-api/0.log" Mar 09 20:16:39 crc kubenswrapper[4750]: I0309 20:16:39.670037 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-79b78f4bf4-klzwx_6116d3b9-23b6-44a7-bf39-3b139a2a771e/barbican-api-log/0.log" Mar 09 20:16:39 crc kubenswrapper[4750]: I0309 20:16:39.713368 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54d9b479b6-f2tns_c44181c7-bb39-45a8-ad11-11f4aab9f8b7/barbican-keystone-listener/0.log" Mar 09 20:16:39 crc kubenswrapper[4750]: I0309 20:16:39.885794 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76cfd56bfc-9l7jh_968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0/barbican-worker/0.log" Mar 09 20:16:39 crc kubenswrapper[4750]: I0309 20:16:39.918811 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54d9b479b6-f2tns_c44181c7-bb39-45a8-ad11-11f4aab9f8b7/barbican-keystone-listener-log/0.log" Mar 09 20:16:39 crc kubenswrapper[4750]: I0309 20:16:39.954741 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76cfd56bfc-9l7jh_968ec7f7-d4c8-4b7f-a3ee-c9e449fc73e0/barbican-worker-log/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.099002 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-jdmk4_edf12ec8-96ec-4d8e-a9f5-bbdd0cdcc4ee/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.189435 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.284782 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_60e48e88-395d-4b6d-9815-70165c7782b5/ceilometer-central-agent/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.327219 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_60e48e88-395d-4b6d-9815-70165c7782b5/ceilometer-notification-agent/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.334817 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_60e48e88-395d-4b6d-9815-70165c7782b5/proxy-httpd/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.422108 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_60e48e88-395d-4b6d-9815-70165c7782b5/sg-core/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.863299 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_dbe58b04-c73c-4292-825f-1e6b466ad54f/cinder-backup/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.933089 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_af01db09-e910-4f84-87d7-d8565dd13633/cinder-api-log/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.970494 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_af01db09-e910-4f84-87d7-d8565dd13633/cinder-api/0.log" Mar 09 20:16:40 crc kubenswrapper[4750]: I0309 20:16:40.972219 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_dbe58b04-c73c-4292-825f-1e6b466ad54f/probe/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.106781 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" exitCode=1 Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.106833 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400"} Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.106874 4750 scope.go:117] "RemoveContainer" containerID="3de89ea73d5af7a2a2e9322545a3df3feb95fe190dd167af00807ec175f3e375" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.107741 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:16:41 crc kubenswrapper[4750]: E0309 20:16:41.108087 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.123027 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e2488a7a-65a2-442f-81bf-5c74ce640dc4/cinder-scheduler/17.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.190008 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.192155 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e2488a7a-65a2-442f-81bf-5c74ce640dc4/probe/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.407602 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_e27746ea-410c-45a2-91f5-58633d7c7747/probe/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.435566 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_e27746ea-410c-45a2-91f5-58633d7c7747/cinder-volume/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.700828 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_a22fae28-039f-4b62-b889-f53da0ef827b/probe/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.701313 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_a22fae28-039f-4b62-b889-f53da0ef827b/cinder-volume/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.745268 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-7t27p_ea9561e1-8a07-4d8a-9726-d1e06b93af82/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.892580 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-cmlf5_ecc06a4b-b3ce-4a96-baef-c1cedf0e45c2/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:41 crc kubenswrapper[4750]: I0309 20:16:41.954601 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69685b9687-h4kfk_19a73cb2-d4f6-41fa-a9ee-8a949bd75c73/init/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.118883 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:16:42 crc kubenswrapper[4750]: E0309 20:16:42.119566 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.169917 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69685b9687-h4kfk_19a73cb2-d4f6-41fa-a9ee-8a949bd75c73/init/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.240166 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9dhzm_825b5a72-cb74-421a-8ca9-103d99691830/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.303594 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69685b9687-h4kfk_19a73cb2-d4f6-41fa-a9ee-8a949bd75c73/dnsmasq-dns/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.443090 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3e04ac5d-19d1-4a07-846e-22852e6a0ca6/glance-httpd/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.462087 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3e04ac5d-19d1-4a07-846e-22852e6a0ca6/glance-log/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.691216 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8b839eed-70cc-4143-880a-52591ffb61c5/glance-httpd/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.705762 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8b839eed-70cc-4143-880a-52591ffb61c5/glance-log/0.log" Mar 09 20:16:42 crc kubenswrapper[4750]: I0309 20:16:42.897252 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fdd5dd964-x7n8q_0509f0f1-226d-4626-998d-b9065a8a634b/horizon/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.072618 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-zjw2n_d46d43ed-f599-4bdd-bacd-3e3f8bb60fca/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.205148 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-xk6t9_316b2da2-4ec6-4c70-83ab-4c6b5e330032/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.488300 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fdd5dd964-x7n8q_0509f0f1-226d-4626-998d-b9065a8a634b/horizon-log/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.617825 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6594b67854-xmvs4_f1171881-c394-4042-b092-a6935ed182df/keystone-api/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.676715 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29551441-8sgzm_5137a335-d7e5-4aff-bb01-19eb8d5600d6/keystone-cron/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.693413 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29551381-lcv2d_1d707807-4c3a-4eeb-b9c5-8c785bdf3cd1/keystone-cron/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.811045 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_47b019da-8326-4a33-80f4-257dd5517e92/kube-state-metrics/0.log" Mar 09 20:16:43 crc kubenswrapper[4750]: I0309 20:16:43.924950 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-827lf_bcbdff51-5cc0-4011-a1ae-a260ee565ef7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:44 crc kubenswrapper[4750]: I0309 20:16:44.131828 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cd6d4c84f-zrhsg_cd28fe12-963a-4e2a-be20-6e58f8034a6f/neutron-api/0.log" Mar 09 20:16:44 crc kubenswrapper[4750]: I0309 20:16:44.195356 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cd6d4c84f-zrhsg_cd28fe12-963a-4e2a-be20-6e58f8034a6f/neutron-httpd/0.log" Mar 09 20:16:44 crc kubenswrapper[4750]: I0309 20:16:44.642210 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-f7pqn_20df7645-4b55-487a-85c1-cb23099b9527/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:44 crc kubenswrapper[4750]: I0309 20:16:44.691988 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_notifications-rabbitmq-server-0_3926781f-b03f-45a7-82e5-10abf0c16c1e/setup-container/0.log" Mar 09 20:16:44 crc kubenswrapper[4750]: I0309 20:16:44.942504 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_notifications-rabbitmq-server-0_3926781f-b03f-45a7-82e5-10abf0c16c1e/setup-container/0.log" Mar 09 20:16:45 crc kubenswrapper[4750]: I0309 20:16:45.038505 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_notifications-rabbitmq-server-0_3926781f-b03f-45a7-82e5-10abf0c16c1e/rabbitmq/0.log" Mar 09 20:16:45 crc kubenswrapper[4750]: I0309 20:16:45.189335 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:16:45 crc kubenswrapper[4750]: I0309 20:16:45.190244 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:16:45 crc kubenswrapper[4750]: E0309 20:16:45.190498 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:16:45 crc kubenswrapper[4750]: I0309 20:16:45.501865 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_50d1b356-0b06-47eb-9fdb-6674038e109d/nova-api-log/0.log" Mar 09 20:16:45 crc kubenswrapper[4750]: I0309 20:16:45.584391 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e0810e68-2041-4716-bcf1-97bd6c2c44aa/nova-cell0-conductor-conductor/0.log" Mar 09 20:16:45 crc kubenswrapper[4750]: I0309 20:16:45.860857 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_78014cb6-5fde-4f65-990e-9c46da7a8578/nova-cell1-conductor-conductor/0.log" Mar 09 20:16:45 crc kubenswrapper[4750]: I0309 20:16:45.948255 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_50d1b356-0b06-47eb-9fdb-6674038e109d/nova-api-api/0.log" Mar 09 20:16:46 crc kubenswrapper[4750]: I0309 20:16:46.053829 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_bfba6047-48a3-4895-90c1-fadceadccba3/nova-cell1-novncproxy-novncproxy/0.log" Mar 09 20:16:46 crc kubenswrapper[4750]: I0309 20:16:46.279411 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-xlzzm_f8f28da2-4627-4099-8838-4ac2b22c2209/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:46 crc kubenswrapper[4750]: I0309 20:16:46.501277 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5f24d405-c297-4d1c-a05c-570e9d532574/nova-metadata-log/0.log" Mar 09 20:16:46 crc kubenswrapper[4750]: I0309 20:16:46.721242 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d4cb5c89-cbc8-469d-afe9-45100b0867e0/nova-scheduler-scheduler/0.log" Mar 09 20:16:46 crc kubenswrapper[4750]: I0309 20:16:46.793969 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7fd7273f-010c-48a0-9ecc-bd80e0bff239/mysql-bootstrap/0.log" Mar 09 20:16:47 crc kubenswrapper[4750]: I0309 20:16:47.129137 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7fd7273f-010c-48a0-9ecc-bd80e0bff239/mysql-bootstrap/0.log" Mar 09 20:16:47 crc kubenswrapper[4750]: I0309 20:16:47.446509 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7fd7273f-010c-48a0-9ecc-bd80e0bff239/galera/0.log" Mar 09 20:16:47 crc kubenswrapper[4750]: I0309 20:16:47.554520 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_302428aa-7eba-4f8b-8868-1e8e883d38c2/mysql-bootstrap/0.log" Mar 09 20:16:47 crc kubenswrapper[4750]: I0309 20:16:47.757232 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_302428aa-7eba-4f8b-8868-1e8e883d38c2/mysql-bootstrap/0.log" Mar 09 20:16:47 crc kubenswrapper[4750]: I0309 20:16:47.798351 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_302428aa-7eba-4f8b-8868-1e8e883d38c2/galera/0.log" Mar 09 20:16:47 crc kubenswrapper[4750]: I0309 20:16:47.982085 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_be47c3f6-61fd-45da-b720-e1df25dd5873/openstackclient/0.log" Mar 09 20:16:48 crc kubenswrapper[4750]: I0309 20:16:48.083892 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nz6xk_ab758bdb-9d1f-4941-8c95-84cc4c867f09/openstack-network-exporter/0.log" Mar 09 20:16:48 crc kubenswrapper[4750]: I0309 20:16:48.284341 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-n5qz8_5e9fd826-3829-4379-9717-1b76ef376a39/ovn-controller/0.log" Mar 09 20:16:48 crc kubenswrapper[4750]: I0309 20:16:48.445226 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6dgcc_89841873-12eb-4add-bf8a-061b624c4b3a/ovsdb-server-init/0.log" Mar 09 20:16:48 crc kubenswrapper[4750]: I0309 20:16:48.681098 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6dgcc_89841873-12eb-4add-bf8a-061b624c4b3a/ovsdb-server-init/0.log" Mar 09 20:16:48 crc kubenswrapper[4750]: I0309 20:16:48.684545 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6dgcc_89841873-12eb-4add-bf8a-061b624c4b3a/ovsdb-server/0.log" Mar 09 20:16:48 crc kubenswrapper[4750]: I0309 20:16:48.743151 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6dgcc_89841873-12eb-4add-bf8a-061b624c4b3a/ovs-vswitchd/0.log" Mar 09 20:16:48 crc kubenswrapper[4750]: I0309 20:16:48.936966 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6l9mf_39d323a7-f34e-404f-ae88-9a3322950f85/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.013662 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5f24d405-c297-4d1c-a05c-570e9d532574/nova-metadata-metadata/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.165056 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2fe00ea1-72f0-4ba3-8096-11e8ec17c733/ovn-northd/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.200399 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2fe00ea1-72f0-4ba3-8096-11e8ec17c733/openstack-network-exporter/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.355814 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f059f39a-1c7f-46af-be90-4095105df733/openstack-network-exporter/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.449421 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f059f39a-1c7f-46af-be90-4095105df733/ovsdbserver-nb/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.505565 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0dbcbd0d-a357-41da-9e5c-6672e4e326fe/openstack-network-exporter/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.595989 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0dbcbd0d-a357-41da-9e5c-6672e4e326fe/ovsdbserver-sb/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.872685 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-64cbf4d6d8-cn7dn_92279388-e576-45e4-a4e6-028db5103301/placement-api/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.892182 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-64cbf4d6d8-cn7dn_92279388-e576-45e4-a4e6-028db5103301/placement-log/0.log" Mar 09 20:16:49 crc kubenswrapper[4750]: I0309 20:16:49.941403 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6ca49786-3511-4719-90de-963bafdd8429/init-config-reloader/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.095256 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6ca49786-3511-4719-90de-963bafdd8429/init-config-reloader/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.114900 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6ca49786-3511-4719-90de-963bafdd8429/config-reloader/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.128292 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6ca49786-3511-4719-90de-963bafdd8429/prometheus/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.165992 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6ca49786-3511-4719-90de-963bafdd8429/thanos-sidecar/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.331788 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b4103fad-b50b-450d-99c0-38a505bd800a/setup-container/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.545829 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b4103fad-b50b-450d-99c0-38a505bd800a/setup-container/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.566598 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b4103fad-b50b-450d-99c0-38a505bd800a/rabbitmq/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.611612 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_14f2384b-45f7-45ca-ac90-429fc2f69bbb/setup-container/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.827950 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_14f2384b-45f7-45ca-ac90-429fc2f69bbb/setup-container/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.901206 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_14f2384b-45f7-45ca-ac90-429fc2f69bbb/rabbitmq/0.log" Mar 09 20:16:50 crc kubenswrapper[4750]: I0309 20:16:50.968560 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xjk64_56a126ce-a44f-449c-b0bc-05da61503d74/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:51 crc kubenswrapper[4750]: I0309 20:16:51.344754 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-tdfct_c4dc3075-1b7c-4076-9beb-f7e5471c2dbe/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:51 crc kubenswrapper[4750]: I0309 20:16:51.395177 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jd2pw_dc1ea8e4-4151-4d4c-b2df-db2976ef8ff6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:51 crc kubenswrapper[4750]: I0309 20:16:51.660107 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-96w2s_50969409-de1c-43a2-bf66-60da87e8bcb4/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:51 crc kubenswrapper[4750]: I0309 20:16:51.666119 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-c5nfl_5679eb0f-6cb5-4d30-8428-19bcf4dc13ed/ssh-known-hosts-edpm-deployment/0.log" Mar 09 20:16:51 crc kubenswrapper[4750]: I0309 20:16:51.956811 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-54cf6684f5-bz5sx_a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96/proxy-server/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.065716 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bjl5g_64b40944-803f-45ca-9d47-86c4250de34b/swift-ring-rebalance/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.165825 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-54cf6684f5-bz5sx_a6c283c4-989e-4acc-ae9d-b7dbdc3d2b96/proxy-httpd/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.184752 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/account-auditor/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.289709 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/account-reaper/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.383457 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/account-server/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.461206 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/account-replicator/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.488607 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/container-auditor/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.587840 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/container-replicator/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.631427 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/container-server/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.664656 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/container-updater/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.734299 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/object-auditor/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.831898 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/object-server/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.857529 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/object-expirer/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.859220 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/object-replicator/0.log" Mar 09 20:16:52 crc kubenswrapper[4750]: I0309 20:16:52.962241 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/object-updater/0.log" Mar 09 20:16:53 crc kubenswrapper[4750]: I0309 20:16:53.071163 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/rsync/0.log" Mar 09 20:16:53 crc kubenswrapper[4750]: I0309 20:16:53.087347 4750 scope.go:117] "RemoveContainer" containerID="553e65245cc8b523056bc0fd43895fc062ed9035198aa566a3c729ae1daecaf8" Mar 09 20:16:53 crc kubenswrapper[4750]: I0309 20:16:53.171269 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_cd12d247-7fd6-468d-8450-c395c4ee57ac/swift-recon-cron/0.log" Mar 09 20:16:53 crc kubenswrapper[4750]: I0309 20:16:53.287730 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2mknd_c0cdad8f-0238-4da8-94ae-39b139b83613/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:53 crc kubenswrapper[4750]: I0309 20:16:53.368326 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-fkvp5_dff3602d-ae54-49a1-8d52-b7eb251a680e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 09 20:16:53 crc kubenswrapper[4750]: I0309 20:16:53.735710 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_664aaf56-bbea-4a12-95d5-ab56a978d608/watcher-api-log/0.log" Mar 09 20:16:53 crc kubenswrapper[4750]: I0309 20:16:53.892436 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_d7f9e82f-cb7a-425d-b942-b41d9c167784/watcher-applier/0.log" Mar 09 20:16:54 crc kubenswrapper[4750]: I0309 20:16:54.554756 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_ab65e149-9e3c-451f-82a4-e0fbe5127ac1/watcher-decision-engine/0.log" Mar 09 20:16:57 crc kubenswrapper[4750]: I0309 20:16:57.562804 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_664aaf56-bbea-4a12-95d5-ab56a978d608/watcher-api/0.log" Mar 09 20:16:58 crc kubenswrapper[4750]: I0309 20:16:58.373061 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:16:58 crc kubenswrapper[4750]: E0309 20:16:58.373929 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:17:07 crc kubenswrapper[4750]: I0309 20:17:07.609912 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_1b310231-1c53-4831-aa00-e0f9597ce6e2/memcached/0.log" Mar 09 20:17:11 crc kubenswrapper[4750]: I0309 20:17:11.373707 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:17:11 crc kubenswrapper[4750]: E0309 20:17:11.374182 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:17:21 crc kubenswrapper[4750]: I0309 20:17:21.743453 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:17:21 crc kubenswrapper[4750]: I0309 20:17:21.744231 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:17:23 crc kubenswrapper[4750]: I0309 20:17:23.573541 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt_f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d/util/0.log" Mar 09 20:17:23 crc kubenswrapper[4750]: I0309 20:17:23.696971 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt_f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d/util/0.log" Mar 09 20:17:23 crc kubenswrapper[4750]: I0309 20:17:23.754797 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt_f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d/pull/0.log" Mar 09 20:17:23 crc kubenswrapper[4750]: I0309 20:17:23.754841 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt_f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d/pull/0.log" Mar 09 20:17:23 crc kubenswrapper[4750]: I0309 20:17:23.925349 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt_f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d/util/0.log" Mar 09 20:17:23 crc kubenswrapper[4750]: I0309 20:17:23.967226 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt_f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d/extract/0.log" Mar 09 20:17:23 crc kubenswrapper[4750]: I0309 20:17:23.971906 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5eea759a77a44b7d379d7a90e28614a746a8848e17a3c9b1bbf53168bfvp7bt_f52fa7ee-86a6-4ca2-95b4-d1b7a10ec23d/pull/0.log" Mar 09 20:17:24 crc kubenswrapper[4750]: I0309 20:17:24.399062 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-5d87c9d997-czkhf_bf342fef-8819-4328-9551-13d21041bff6/manager/0.log" Mar 09 20:17:25 crc kubenswrapper[4750]: I0309 20:17:25.203736 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-64db6967f8-wl9k2_c7f102b3-0989-4582-91d1-49a22c261371/manager/0.log" Mar 09 20:17:25 crc kubenswrapper[4750]: I0309 20:17:25.226920 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-cf99c678f-xrjb7_434cb9b7-d90c-4081-8232-004bbd4cb8b2/manager/0.log" Mar 09 20:17:25 crc kubenswrapper[4750]: I0309 20:17:25.487860 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-78bc7f9bd9-tkcr6_f4e026ba-f7dd-4c14-b09e-e00e6cfae20d/manager/0.log" Mar 09 20:17:26 crc kubenswrapper[4750]: I0309 20:17:26.095660 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-545456dc4-lmj58_3f316fad-80d2-46fb-ab97-d3f2b0559387/manager/0.log" Mar 09 20:17:26 crc kubenswrapper[4750]: I0309 20:17:26.319725 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-f7fcc58b9-8pvfn_6f320add-8d8a-49c6-a8bd-c286772ee907/manager/0.log" Mar 09 20:17:26 crc kubenswrapper[4750]: I0309 20:17:26.373317 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:17:26 crc kubenswrapper[4750]: E0309 20:17:26.373556 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:17:26 crc kubenswrapper[4750]: I0309 20:17:26.650240 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7c789f89c6-j799f_a4506343-59e4-4442-8894-fdb158c82316/manager/0.log" Mar 09 20:17:26 crc kubenswrapper[4750]: I0309 20:17:26.849240 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-wrkj4_e671bb3b-ca69-4229-9e9b-e823bfcf2f92/manager/0.log" Mar 09 20:17:27 crc kubenswrapper[4750]: I0309 20:17:27.155032 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7b6bfb6475-znrqs_d8208475-d487-47ee-8f03-fa4379d3c917/manager/0.log" Mar 09 20:17:27 crc kubenswrapper[4750]: I0309 20:17:27.374220 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-86r5x_607af51a-435c-4f0a-99c2-12c3cf6b2213/manager/0.log" Mar 09 20:17:27 crc kubenswrapper[4750]: I0309 20:17:27.469305 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54688575f-n8ldk_1dc593d7-29a0-4a18-aecc-81f22e19e9b9/manager/0.log" Mar 09 20:17:27 crc kubenswrapper[4750]: I0309 20:17:27.700354 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5d86c7ddb7-dnt7g_c42f78da-7508-4f90-bee6-c05e5cbd0f86/manager/0.log" Mar 09 20:17:27 crc kubenswrapper[4750]: I0309 20:17:27.747138 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-74b6b5dc96-qp844_14f86f92-1676-44e4-a56c-e611b59afe30/manager/0.log" Mar 09 20:17:28 crc kubenswrapper[4750]: I0309 20:17:28.002651 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9c5gstw_89799193-9df0-4505-8116-d49a86d33add/manager/0.log" Mar 09 20:17:28 crc kubenswrapper[4750]: I0309 20:17:28.211841 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-787cf98cf6-rb7j4_7fdb116b-f1d9-4e68-ad54-ed2217b6f001/operator/0.log" Mar 09 20:17:28 crc kubenswrapper[4750]: I0309 20:17:28.362809 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-q9qqf_a08a8a31-630e-45ad-a319-c2904fa895b3/registry-server/0.log" Mar 09 20:17:28 crc kubenswrapper[4750]: I0309 20:17:28.540037 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-75684d597f-hj5hw_1507d1a6-11a2-4a35-9c3c-1c360e4c0b32/manager/0.log" Mar 09 20:17:28 crc kubenswrapper[4750]: I0309 20:17:28.766982 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-648564c9fc-2l9js_93ac04dd-4b80-4e20-9113-eadef2054d0e/manager/0.log" Mar 09 20:17:28 crc kubenswrapper[4750]: I0309 20:17:28.826610 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fbjb2_a91098d3-ee59-4f84-ad48-ee7613a8c7da/operator/0.log" Mar 09 20:17:29 crc kubenswrapper[4750]: I0309 20:17:29.026879 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9b9ff9f4d-skdbr_ccf58e0b-b6b3-443a-b215-087bb55da05c/manager/0.log" Mar 09 20:17:29 crc kubenswrapper[4750]: I0309 20:17:29.314210 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-55b5ff4dbb-z96st_8c687b21-5077-4d33-af70-07b01956d904/manager/0.log" Mar 09 20:17:29 crc kubenswrapper[4750]: I0309 20:17:29.521360 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5fdb694969-rvpms_e403dea7-17d0-40db-9b29-e50f33175503/manager/0.log" Mar 09 20:17:29 crc kubenswrapper[4750]: I0309 20:17:29.683729 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-668c5c65dc-rt6bl_6bd1943b-5152-4d4e-9fa9-8c2f8ea42072/manager/0.log" Mar 09 20:17:30 crc kubenswrapper[4750]: I0309 20:17:30.335948 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64797568c9-p55cp_874a050d-1e92-4087-a655-b6fdb66a47e9/manager/0.log" Mar 09 20:17:36 crc kubenswrapper[4750]: I0309 20:17:36.300508 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6db6876945-qdwbn_9b0d0f24-6c38-4ccb-8e4d-a0784c006dd9/manager/0.log" Mar 09 20:17:37 crc kubenswrapper[4750]: I0309 20:17:37.373412 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:17:37 crc kubenswrapper[4750]: E0309 20:17:37.374317 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:17:51 crc kubenswrapper[4750]: I0309 20:17:51.373788 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:17:51 crc kubenswrapper[4750]: E0309 20:17:51.374807 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:17:51 crc kubenswrapper[4750]: I0309 20:17:51.742936 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:17:51 crc kubenswrapper[4750]: I0309 20:17:51.743207 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:17:51 crc kubenswrapper[4750]: I0309 20:17:51.873864 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-p5d9d_dac5b814-763c-42f7-a499-7a6afe1787b9/control-plane-machine-set-operator/0.log" Mar 09 20:17:52 crc kubenswrapper[4750]: I0309 20:17:52.048772 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5c9r_027f6e21-69f1-4111-85ec-f4f7486da3a5/kube-rbac-proxy/0.log" Mar 09 20:17:52 crc kubenswrapper[4750]: I0309 20:17:52.061964 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5c9r_027f6e21-69f1-4111-85ec-f4f7486da3a5/machine-api-operator/0.log" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.148611 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551458-2pm68"] Mar 09 20:18:00 crc kubenswrapper[4750]: E0309 20:18:00.149592 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756bf273-1cee-409f-9c91-b45473160359" containerName="oc" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.149607 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="756bf273-1cee-409f-9c91-b45473160359" containerName="oc" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.149858 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="756bf273-1cee-409f-9c91-b45473160359" containerName="oc" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.150532 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551458-2pm68" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.156100 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.157134 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.157531 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.173323 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551458-2pm68"] Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.196062 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbqf2\" (UniqueName: \"kubernetes.io/projected/6f078a8e-15b4-4a78-8566-0d650ffcbf4c-kube-api-access-hbqf2\") pod \"auto-csr-approver-29551458-2pm68\" (UID: \"6f078a8e-15b4-4a78-8566-0d650ffcbf4c\") " pod="openshift-infra/auto-csr-approver-29551458-2pm68" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.298458 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbqf2\" (UniqueName: \"kubernetes.io/projected/6f078a8e-15b4-4a78-8566-0d650ffcbf4c-kube-api-access-hbqf2\") pod \"auto-csr-approver-29551458-2pm68\" (UID: \"6f078a8e-15b4-4a78-8566-0d650ffcbf4c\") " pod="openshift-infra/auto-csr-approver-29551458-2pm68" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.317865 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbqf2\" (UniqueName: \"kubernetes.io/projected/6f078a8e-15b4-4a78-8566-0d650ffcbf4c-kube-api-access-hbqf2\") pod \"auto-csr-approver-29551458-2pm68\" (UID: \"6f078a8e-15b4-4a78-8566-0d650ffcbf4c\") " pod="openshift-infra/auto-csr-approver-29551458-2pm68" Mar 09 20:18:00 crc kubenswrapper[4750]: I0309 20:18:00.474708 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551458-2pm68" Mar 09 20:18:01 crc kubenswrapper[4750]: I0309 20:18:01.061927 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551458-2pm68"] Mar 09 20:18:01 crc kubenswrapper[4750]: I0309 20:18:01.072978 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 20:18:01 crc kubenswrapper[4750]: I0309 20:18:01.872675 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551458-2pm68" event={"ID":"6f078a8e-15b4-4a78-8566-0d650ffcbf4c","Type":"ContainerStarted","Data":"2db4a1a4a07a3c4702f45ac05fcbc2597717df79922124cea4eb1b45d7531738"} Mar 09 20:18:02 crc kubenswrapper[4750]: I0309 20:18:02.883409 4750 generic.go:334] "Generic (PLEG): container finished" podID="6f078a8e-15b4-4a78-8566-0d650ffcbf4c" containerID="af2fec232aec7c00cf3be0a2b8c76305b47b2d45b7f62a9249b7bc1793f0946b" exitCode=0 Mar 09 20:18:02 crc kubenswrapper[4750]: I0309 20:18:02.883453 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551458-2pm68" event={"ID":"6f078a8e-15b4-4a78-8566-0d650ffcbf4c","Type":"ContainerDied","Data":"af2fec232aec7c00cf3be0a2b8c76305b47b2d45b7f62a9249b7bc1793f0946b"} Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.405912 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9mw"] Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.412282 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.420729 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9mw"] Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.582053 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-catalog-content\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.582397 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xq5n\" (UniqueName: \"kubernetes.io/projected/64301215-274f-4356-99f0-b4ad5a3e9b8c-kube-api-access-6xq5n\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.582875 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-utilities\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.684572 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-catalog-content\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.684639 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xq5n\" (UniqueName: \"kubernetes.io/projected/64301215-274f-4356-99f0-b4ad5a3e9b8c-kube-api-access-6xq5n\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.684761 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-utilities\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.685107 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-catalog-content\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.685281 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-utilities\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.714486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xq5n\" (UniqueName: \"kubernetes.io/projected/64301215-274f-4356-99f0-b4ad5a3e9b8c-kube-api-access-6xq5n\") pod \"redhat-marketplace-rt9mw\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:03 crc kubenswrapper[4750]: I0309 20:18:03.795416 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.319175 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9mw"] Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.369811 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551458-2pm68" Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.502455 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbqf2\" (UniqueName: \"kubernetes.io/projected/6f078a8e-15b4-4a78-8566-0d650ffcbf4c-kube-api-access-hbqf2\") pod \"6f078a8e-15b4-4a78-8566-0d650ffcbf4c\" (UID: \"6f078a8e-15b4-4a78-8566-0d650ffcbf4c\") " Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.508815 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f078a8e-15b4-4a78-8566-0d650ffcbf4c-kube-api-access-hbqf2" (OuterVolumeSpecName: "kube-api-access-hbqf2") pod "6f078a8e-15b4-4a78-8566-0d650ffcbf4c" (UID: "6f078a8e-15b4-4a78-8566-0d650ffcbf4c"). InnerVolumeSpecName "kube-api-access-hbqf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.605710 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbqf2\" (UniqueName: \"kubernetes.io/projected/6f078a8e-15b4-4a78-8566-0d650ffcbf4c-kube-api-access-hbqf2\") on node \"crc\" DevicePath \"\"" Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.916310 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551458-2pm68" event={"ID":"6f078a8e-15b4-4a78-8566-0d650ffcbf4c","Type":"ContainerDied","Data":"2db4a1a4a07a3c4702f45ac05fcbc2597717df79922124cea4eb1b45d7531738"} Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.916665 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2db4a1a4a07a3c4702f45ac05fcbc2597717df79922124cea4eb1b45d7531738" Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.916847 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551458-2pm68" Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.920564 4750 generic.go:334] "Generic (PLEG): container finished" podID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerID="0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb" exitCode=0 Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.920623 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9mw" event={"ID":"64301215-274f-4356-99f0-b4ad5a3e9b8c","Type":"ContainerDied","Data":"0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb"} Mar 09 20:18:04 crc kubenswrapper[4750]: I0309 20:18:04.920673 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9mw" event={"ID":"64301215-274f-4356-99f0-b4ad5a3e9b8c","Type":"ContainerStarted","Data":"6c7ebce7a343e16a111c344e58e0c54fcc9f227b6ed9b0db71f7d1ce432bc5a0"} Mar 09 20:18:05 crc kubenswrapper[4750]: I0309 20:18:05.373013 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:18:05 crc kubenswrapper[4750]: E0309 20:18:05.373501 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:18:05 crc kubenswrapper[4750]: I0309 20:18:05.449876 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551452-nt87s"] Mar 09 20:18:05 crc kubenswrapper[4750]: I0309 20:18:05.465471 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551452-nt87s"] Mar 09 20:18:05 crc kubenswrapper[4750]: I0309 20:18:05.845308 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-xxfrp_d4fb6a3f-ccb3-4df8-b8f7-2d3b3295448c/cert-manager-controller/0.log" Mar 09 20:18:05 crc kubenswrapper[4750]: I0309 20:18:05.937495 4750 generic.go:334] "Generic (PLEG): container finished" podID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerID="37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436" exitCode=0 Mar 09 20:18:05 crc kubenswrapper[4750]: I0309 20:18:05.937570 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9mw" event={"ID":"64301215-274f-4356-99f0-b4ad5a3e9b8c","Type":"ContainerDied","Data":"37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436"} Mar 09 20:18:05 crc kubenswrapper[4750]: I0309 20:18:05.999911 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-2k6r6_45e5d4bf-93bf-4bea-9ee4-656c3c4f8077/cert-manager-cainjector/0.log" Mar 09 20:18:06 crc kubenswrapper[4750]: I0309 20:18:06.094672 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-8pf5j_b7eaca9e-324d-4934-bcac-d8a6034c94f6/cert-manager-webhook/0.log" Mar 09 20:18:06 crc kubenswrapper[4750]: I0309 20:18:06.947835 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9mw" event={"ID":"64301215-274f-4356-99f0-b4ad5a3e9b8c","Type":"ContainerStarted","Data":"a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9"} Mar 09 20:18:06 crc kubenswrapper[4750]: I0309 20:18:06.966832 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rt9mw" podStartSLOduration=2.533839983 podStartE2EDuration="3.96681009s" podCreationTimestamp="2026-03-09 20:18:03 +0000 UTC" firstStartedPulling="2026-03-09 20:18:04.922708161 +0000 UTC m=+6766.265180579" lastFinishedPulling="2026-03-09 20:18:06.355678288 +0000 UTC m=+6767.698150686" observedRunningTime="2026-03-09 20:18:06.962957796 +0000 UTC m=+6768.305430194" watchObservedRunningTime="2026-03-09 20:18:06.96681009 +0000 UTC m=+6768.309282488" Mar 09 20:18:07 crc kubenswrapper[4750]: I0309 20:18:07.383270 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c136bd51-ed7b-4b3b-988e-420eadff782c" path="/var/lib/kubelet/pods/c136bd51-ed7b-4b3b-988e-420eadff782c/volumes" Mar 09 20:18:13 crc kubenswrapper[4750]: I0309 20:18:13.796497 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:13 crc kubenswrapper[4750]: I0309 20:18:13.797028 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:13 crc kubenswrapper[4750]: I0309 20:18:13.851389 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:14 crc kubenswrapper[4750]: I0309 20:18:14.058238 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:14 crc kubenswrapper[4750]: I0309 20:18:14.109548 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9mw"] Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.033589 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rt9mw" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="registry-server" containerID="cri-o://a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9" gracePeriod=2 Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.374321 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:18:16 crc kubenswrapper[4750]: E0309 20:18:16.374811 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.486363 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.650028 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xq5n\" (UniqueName: \"kubernetes.io/projected/64301215-274f-4356-99f0-b4ad5a3e9b8c-kube-api-access-6xq5n\") pod \"64301215-274f-4356-99f0-b4ad5a3e9b8c\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.650110 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-catalog-content\") pod \"64301215-274f-4356-99f0-b4ad5a3e9b8c\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.650148 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-utilities\") pod \"64301215-274f-4356-99f0-b4ad5a3e9b8c\" (UID: \"64301215-274f-4356-99f0-b4ad5a3e9b8c\") " Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.651755 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-utilities" (OuterVolumeSpecName: "utilities") pod "64301215-274f-4356-99f0-b4ad5a3e9b8c" (UID: "64301215-274f-4356-99f0-b4ad5a3e9b8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.657652 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64301215-274f-4356-99f0-b4ad5a3e9b8c-kube-api-access-6xq5n" (OuterVolumeSpecName: "kube-api-access-6xq5n") pod "64301215-274f-4356-99f0-b4ad5a3e9b8c" (UID: "64301215-274f-4356-99f0-b4ad5a3e9b8c"). InnerVolumeSpecName "kube-api-access-6xq5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.689245 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64301215-274f-4356-99f0-b4ad5a3e9b8c" (UID: "64301215-274f-4356-99f0-b4ad5a3e9b8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.753152 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xq5n\" (UniqueName: \"kubernetes.io/projected/64301215-274f-4356-99f0-b4ad5a3e9b8c-kube-api-access-6xq5n\") on node \"crc\" DevicePath \"\"" Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.753197 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:18:16 crc kubenswrapper[4750]: I0309 20:18:16.753210 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64301215-274f-4356-99f0-b4ad5a3e9b8c-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.057847 4750 generic.go:334] "Generic (PLEG): container finished" podID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerID="a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9" exitCode=0 Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.057918 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9mw" event={"ID":"64301215-274f-4356-99f0-b4ad5a3e9b8c","Type":"ContainerDied","Data":"a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9"} Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.057951 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rt9mw" event={"ID":"64301215-274f-4356-99f0-b4ad5a3e9b8c","Type":"ContainerDied","Data":"6c7ebce7a343e16a111c344e58e0c54fcc9f227b6ed9b0db71f7d1ce432bc5a0"} Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.057973 4750 scope.go:117] "RemoveContainer" containerID="a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.058438 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rt9mw" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.097213 4750 scope.go:117] "RemoveContainer" containerID="37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.098115 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9mw"] Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.110602 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rt9mw"] Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.120703 4750 scope.go:117] "RemoveContainer" containerID="0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.184687 4750 scope.go:117] "RemoveContainer" containerID="a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9" Mar 09 20:18:17 crc kubenswrapper[4750]: E0309 20:18:17.186351 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9\": container with ID starting with a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9 not found: ID does not exist" containerID="a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.186407 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9"} err="failed to get container status \"a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9\": rpc error: code = NotFound desc = could not find container \"a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9\": container with ID starting with a2ad57ff5ef41e62a18db0146135f65c5406e78d121d9265a3728e1b862016c9 not found: ID does not exist" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.186442 4750 scope.go:117] "RemoveContainer" containerID="37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436" Mar 09 20:18:17 crc kubenswrapper[4750]: E0309 20:18:17.186996 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436\": container with ID starting with 37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436 not found: ID does not exist" containerID="37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.187039 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436"} err="failed to get container status \"37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436\": rpc error: code = NotFound desc = could not find container \"37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436\": container with ID starting with 37bac45e92ba73ce90a6f0ac9eda59d4a89d0442144731f97ad6441746447436 not found: ID does not exist" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.187071 4750 scope.go:117] "RemoveContainer" containerID="0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb" Mar 09 20:18:17 crc kubenswrapper[4750]: E0309 20:18:17.187383 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb\": container with ID starting with 0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb not found: ID does not exist" containerID="0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.187429 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb"} err="failed to get container status \"0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb\": rpc error: code = NotFound desc = could not find container \"0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb\": container with ID starting with 0e4f8b51e39a988d6b455baa790f25c431f8684f26e876c0439a9de6a9d309cb not found: ID does not exist" Mar 09 20:18:17 crc kubenswrapper[4750]: I0309 20:18:17.383751 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" path="/var/lib/kubelet/pods/64301215-274f-4356-99f0-b4ad5a3e9b8c/volumes" Mar 09 20:18:19 crc kubenswrapper[4750]: I0309 20:18:19.414303 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-fg9ks_b3d9d77f-c39e-4e33-a33b-b75a6a9615ef/nmstate-console-plugin/0.log" Mar 09 20:18:19 crc kubenswrapper[4750]: I0309 20:18:19.596149 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-b5lff_b66128f4-2824-4e2a-93ec-4a59f313cb8d/nmstate-handler/0.log" Mar 09 20:18:19 crc kubenswrapper[4750]: I0309 20:18:19.640444 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-xzqjf_5eee205a-209f-4f39-8341-e5bfb1da1374/kube-rbac-proxy/0.log" Mar 09 20:18:19 crc kubenswrapper[4750]: I0309 20:18:19.681553 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-xzqjf_5eee205a-209f-4f39-8341-e5bfb1da1374/nmstate-metrics/0.log" Mar 09 20:18:19 crc kubenswrapper[4750]: I0309 20:18:19.787022 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-tk5qj_a7132a9a-02c3-4361-9ccb-c135f416d289/nmstate-operator/0.log" Mar 09 20:18:19 crc kubenswrapper[4750]: I0309 20:18:19.912723 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-xtfm7_73e03f8a-161c-4019-a06f-41e71db3f4d4/nmstate-webhook/0.log" Mar 09 20:18:21 crc kubenswrapper[4750]: I0309 20:18:21.743283 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:18:21 crc kubenswrapper[4750]: I0309 20:18:21.743656 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:18:21 crc kubenswrapper[4750]: I0309 20:18:21.743697 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 20:18:21 crc kubenswrapper[4750]: I0309 20:18:21.744372 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 20:18:21 crc kubenswrapper[4750]: I0309 20:18:21.744418 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" gracePeriod=600 Mar 09 20:18:21 crc kubenswrapper[4750]: E0309 20:18:21.886736 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:18:22 crc kubenswrapper[4750]: I0309 20:18:22.105585 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" exitCode=0 Mar 09 20:18:22 crc kubenswrapper[4750]: I0309 20:18:22.105671 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8"} Mar 09 20:18:22 crc kubenswrapper[4750]: I0309 20:18:22.105715 4750 scope.go:117] "RemoveContainer" containerID="1aee94e71f7a10634c557910d5c09ce69ff26474679f02e5aab3816f75d7901e" Mar 09 20:18:22 crc kubenswrapper[4750]: I0309 20:18:22.106575 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:18:22 crc kubenswrapper[4750]: E0309 20:18:22.107051 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:18:27 crc kubenswrapper[4750]: I0309 20:18:27.373023 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:18:27 crc kubenswrapper[4750]: E0309 20:18:27.373595 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:18:34 crc kubenswrapper[4750]: I0309 20:18:34.040901 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-7kjhn_8581d9a9-567b-43cc-919e-92099fcffc93/prometheus-operator/0.log" Mar 09 20:18:34 crc kubenswrapper[4750]: I0309 20:18:34.246317 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz_f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86/prometheus-operator-admission-webhook/0.log" Mar 09 20:18:34 crc kubenswrapper[4750]: I0309 20:18:34.352406 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt_e5ff4820-7371-4401-805c-862b76511e97/prometheus-operator-admission-webhook/0.log" Mar 09 20:18:34 crc kubenswrapper[4750]: I0309 20:18:34.454176 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-jbcvg_9b0f552e-2e27-47d5-a0ca-49a041ddb74a/operator/0.log" Mar 09 20:18:34 crc kubenswrapper[4750]: I0309 20:18:34.561353 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-7cmcp_ef54df65-856d-4a28-a4ba-bec1bf76ac07/perses-operator/0.log" Mar 09 20:18:37 crc kubenswrapper[4750]: I0309 20:18:37.375256 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:18:37 crc kubenswrapper[4750]: E0309 20:18:37.375907 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:18:41 crc kubenswrapper[4750]: I0309 20:18:41.374974 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:18:41 crc kubenswrapper[4750]: E0309 20:18:41.376059 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:18:49 crc kubenswrapper[4750]: I0309 20:18:49.578109 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-d9bmt_39f2ab0a-439c-404f-872e-e40e746282d1/kube-rbac-proxy/0.log" Mar 09 20:18:49 crc kubenswrapper[4750]: I0309 20:18:49.634337 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-d9bmt_39f2ab0a-439c-404f-872e-e40e746282d1/controller/0.log" Mar 09 20:18:49 crc kubenswrapper[4750]: I0309 20:18:49.805935 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-frr-files/0.log" Mar 09 20:18:49 crc kubenswrapper[4750]: I0309 20:18:49.958683 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-frr-files/0.log" Mar 09 20:18:49 crc kubenswrapper[4750]: I0309 20:18:49.977012 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-reloader/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.020729 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-metrics/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.030151 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-reloader/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.204623 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-reloader/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.210338 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-metrics/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.226545 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-frr-files/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.243875 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-metrics/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.504008 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/controller/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.522962 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-reloader/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.528669 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-frr-files/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.544474 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/cp-metrics/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.696910 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/frr-metrics/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.732984 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/kube-rbac-proxy/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.759217 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/kube-rbac-proxy-frr/0.log" Mar 09 20:18:50 crc kubenswrapper[4750]: I0309 20:18:50.946374 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/reloader/0.log" Mar 09 20:18:51 crc kubenswrapper[4750]: I0309 20:18:51.009885 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-5s5xj_ef9fe80b-913b-424a-a780-8223aed6eb50/frr-k8s-webhook-server/0.log" Mar 09 20:18:51 crc kubenswrapper[4750]: I0309 20:18:51.245101 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6754f6499c-wbkgp_7878a0a9-d982-4ea1-bc04-e2c2bee5e7fd/manager/0.log" Mar 09 20:18:51 crc kubenswrapper[4750]: I0309 20:18:51.373200 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:18:51 crc kubenswrapper[4750]: E0309 20:18:51.373828 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:18:51 crc kubenswrapper[4750]: I0309 20:18:51.429951 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-749bcdd774-cjqbz_1687d828-b0e8-45ef-8828-00af0b7ea22f/webhook-server/0.log" Mar 09 20:18:51 crc kubenswrapper[4750]: I0309 20:18:51.472285 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bqt4d_4d0613aa-6eb9-43d8-bf29-e63e6b053c7f/kube-rbac-proxy/0.log" Mar 09 20:18:52 crc kubenswrapper[4750]: I0309 20:18:52.127783 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bqt4d_4d0613aa-6eb9-43d8-bf29-e63e6b053c7f/speaker/0.log" Mar 09 20:18:52 crc kubenswrapper[4750]: I0309 20:18:52.794806 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-spjm2_9d818f1c-4170-4157-b621-fb422f11c109/frr/0.log" Mar 09 20:18:53 crc kubenswrapper[4750]: I0309 20:18:53.179068 4750 scope.go:117] "RemoveContainer" containerID="2a4e4e4d9410d1c6f872f7e1b34230ab9bf220fe59359782563ffef9410d2271" Mar 09 20:18:54 crc kubenswrapper[4750]: I0309 20:18:54.374353 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:18:54 crc kubenswrapper[4750]: E0309 20:18:54.375064 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:19:02 crc kubenswrapper[4750]: I0309 20:19:02.374080 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:19:02 crc kubenswrapper[4750]: E0309 20:19:02.375161 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.297277 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6_84a0fae0-78e5-4901-814b-e691c6b1805f/util/0.log" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.493317 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6_84a0fae0-78e5-4901-814b-e691c6b1805f/util/0.log" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.496293 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6_84a0fae0-78e5-4901-814b-e691c6b1805f/pull/0.log" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.513724 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6_84a0fae0-78e5-4901-814b-e691c6b1805f/pull/0.log" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.688812 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6_84a0fae0-78e5-4901-814b-e691c6b1805f/util/0.log" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.694405 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6_84a0fae0-78e5-4901-814b-e691c6b1805f/pull/0.log" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.754445 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82x6jc6_84a0fae0-78e5-4901-814b-e691c6b1805f/extract/0.log" Mar 09 20:19:05 crc kubenswrapper[4750]: I0309 20:19:05.860115 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6_908909ba-f493-489e-834e-863230ee979b/util/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.044441 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6_908909ba-f493-489e-834e-863230ee979b/util/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.076226 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6_908909ba-f493-489e-834e-863230ee979b/pull/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.083087 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6_908909ba-f493-489e-834e-863230ee979b/pull/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.218167 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6_908909ba-f493-489e-834e-863230ee979b/util/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.240247 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6_908909ba-f493-489e-834e-863230ee979b/pull/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.264526 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f088wkc6_908909ba-f493-489e-834e-863230ee979b/extract/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.381181 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5vxm_d497ab4a-bba2-411c-9ffc-7197fc96b0db/extract-utilities/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.565774 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5vxm_d497ab4a-bba2-411c-9ffc-7197fc96b0db/extract-content/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.573235 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5vxm_d497ab4a-bba2-411c-9ffc-7197fc96b0db/extract-utilities/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.584292 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5vxm_d497ab4a-bba2-411c-9ffc-7197fc96b0db/extract-content/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.738162 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5vxm_d497ab4a-bba2-411c-9ffc-7197fc96b0db/extract-content/0.log" Mar 09 20:19:06 crc kubenswrapper[4750]: I0309 20:19:06.792592 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5vxm_d497ab4a-bba2-411c-9ffc-7197fc96b0db/extract-utilities/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.047152 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-55x9g_a2da9cec-0767-4af5-a433-c9a50eeca1fc/extract-utilities/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.055925 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5vxm_d497ab4a-bba2-411c-9ffc-7197fc96b0db/registry-server/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.209013 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-55x9g_a2da9cec-0767-4af5-a433-c9a50eeca1fc/extract-content/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.214302 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-55x9g_a2da9cec-0767-4af5-a433-c9a50eeca1fc/extract-utilities/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.218036 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-55x9g_a2da9cec-0767-4af5-a433-c9a50eeca1fc/extract-content/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.420441 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-55x9g_a2da9cec-0767-4af5-a433-c9a50eeca1fc/extract-utilities/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.481015 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-55x9g_a2da9cec-0767-4af5-a433-c9a50eeca1fc/extract-content/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.633209 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc_bf52966f-af17-425e-a64b-109c9245bcec/util/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.839023 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc_bf52966f-af17-425e-a64b-109c9245bcec/util/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.889989 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc_bf52966f-af17-425e-a64b-109c9245bcec/pull/0.log" Mar 09 20:19:07 crc kubenswrapper[4750]: I0309 20:19:07.911494 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc_bf52966f-af17-425e-a64b-109c9245bcec/pull/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.120075 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc_bf52966f-af17-425e-a64b-109c9245bcec/pull/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.148224 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc_bf52966f-af17-425e-a64b-109c9245bcec/util/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.182449 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4s2hmc_bf52966f-af17-425e-a64b-109c9245bcec/extract/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.358345 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2nkw4_ab532dd6-c38a-4218-bdb3-b4492f1d5f2f/marketplace-operator/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.434391 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-55x9g_a2da9cec-0767-4af5-a433-c9a50eeca1fc/registry-server/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.568592 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-66h4d_06088b9b-c876-4cf1-978e-3aae41970e58/extract-utilities/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.734533 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-66h4d_06088b9b-c876-4cf1-978e-3aae41970e58/extract-content/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.754063 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-66h4d_06088b9b-c876-4cf1-978e-3aae41970e58/extract-content/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.758172 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-66h4d_06088b9b-c876-4cf1-978e-3aae41970e58/extract-utilities/0.log" Mar 09 20:19:08 crc kubenswrapper[4750]: I0309 20:19:08.907187 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-66h4d_06088b9b-c876-4cf1-978e-3aae41970e58/extract-utilities/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:08.999902 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-66h4d_06088b9b-c876-4cf1-978e-3aae41970e58/extract-content/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.030671 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bhdx5_eb5bc9f7-34d2-4180-8633-5deac70271d8/extract-utilities/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.124725 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-66h4d_06088b9b-c876-4cf1-978e-3aae41970e58/registry-server/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.208596 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bhdx5_eb5bc9f7-34d2-4180-8633-5deac70271d8/extract-utilities/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.244416 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bhdx5_eb5bc9f7-34d2-4180-8633-5deac70271d8/extract-content/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.245528 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bhdx5_eb5bc9f7-34d2-4180-8633-5deac70271d8/extract-content/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.384718 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:19:09 crc kubenswrapper[4750]: E0309 20:19:09.385191 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.449770 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bhdx5_eb5bc9f7-34d2-4180-8633-5deac70271d8/extract-content/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.460923 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bhdx5_eb5bc9f7-34d2-4180-8633-5deac70271d8/extract-utilities/0.log" Mar 09 20:19:09 crc kubenswrapper[4750]: I0309 20:19:09.768728 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bhdx5_eb5bc9f7-34d2-4180-8633-5deac70271d8/registry-server/0.log" Mar 09 20:19:16 crc kubenswrapper[4750]: I0309 20:19:16.373857 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:19:16 crc kubenswrapper[4750]: E0309 20:19:16.374777 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:19:21 crc kubenswrapper[4750]: I0309 20:19:21.820561 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7f9c57f9f-8wgwz_f9b6a4c9-be50-4b94-ba18-a9ed1c0c6f86/prometheus-operator-admission-webhook/0.log" Mar 09 20:19:21 crc kubenswrapper[4750]: I0309 20:19:21.821866 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7f9c57f9f-j52xt_e5ff4820-7371-4401-805c-862b76511e97/prometheus-operator-admission-webhook/0.log" Mar 09 20:19:21 crc kubenswrapper[4750]: I0309 20:19:21.893135 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-7kjhn_8581d9a9-567b-43cc-919e-92099fcffc93/prometheus-operator/0.log" Mar 09 20:19:22 crc kubenswrapper[4750]: I0309 20:19:22.012451 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-7cmcp_ef54df65-856d-4a28-a4ba-bec1bf76ac07/perses-operator/0.log" Mar 09 20:19:22 crc kubenswrapper[4750]: I0309 20:19:22.053191 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-jbcvg_9b0f552e-2e27-47d5-a0ca-49a041ddb74a/operator/0.log" Mar 09 20:19:23 crc kubenswrapper[4750]: I0309 20:19:23.373920 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:19:23 crc kubenswrapper[4750]: E0309 20:19:23.374469 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:19:29 crc kubenswrapper[4750]: I0309 20:19:29.383484 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:19:29 crc kubenswrapper[4750]: E0309 20:19:29.384343 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:19:37 crc kubenswrapper[4750]: I0309 20:19:37.373049 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:19:37 crc kubenswrapper[4750]: E0309 20:19:37.373741 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:19:42 crc kubenswrapper[4750]: I0309 20:19:42.373847 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:19:42 crc kubenswrapper[4750]: E0309 20:19:42.374401 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:19:51 crc kubenswrapper[4750]: I0309 20:19:51.374572 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:19:51 crc kubenswrapper[4750]: E0309 20:19:51.375672 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:19:55 crc kubenswrapper[4750]: I0309 20:19:55.374050 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:19:55 crc kubenswrapper[4750]: E0309 20:19:55.374837 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.166523 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551460-4qpgm"] Mar 09 20:20:00 crc kubenswrapper[4750]: E0309 20:20:00.169735 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f078a8e-15b4-4a78-8566-0d650ffcbf4c" containerName="oc" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.169960 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f078a8e-15b4-4a78-8566-0d650ffcbf4c" containerName="oc" Mar 09 20:20:00 crc kubenswrapper[4750]: E0309 20:20:00.171137 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="registry-server" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.171343 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="registry-server" Mar 09 20:20:00 crc kubenswrapper[4750]: E0309 20:20:00.171511 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="extract-utilities" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.171665 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="extract-utilities" Mar 09 20:20:00 crc kubenswrapper[4750]: E0309 20:20:00.171882 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="extract-content" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.172011 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="extract-content" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.172608 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f078a8e-15b4-4a78-8566-0d650ffcbf4c" containerName="oc" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.172875 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="64301215-274f-4356-99f0-b4ad5a3e9b8c" containerName="registry-server" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.174905 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551460-4qpgm" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.178154 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.178356 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.178803 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.180801 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551460-4qpgm"] Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.218607 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzvd6\" (UniqueName: \"kubernetes.io/projected/a876e277-510e-4d8a-9c59-a7517de82dba-kube-api-access-wzvd6\") pod \"auto-csr-approver-29551460-4qpgm\" (UID: \"a876e277-510e-4d8a-9c59-a7517de82dba\") " pod="openshift-infra/auto-csr-approver-29551460-4qpgm" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.319367 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzvd6\" (UniqueName: \"kubernetes.io/projected/a876e277-510e-4d8a-9c59-a7517de82dba-kube-api-access-wzvd6\") pod \"auto-csr-approver-29551460-4qpgm\" (UID: \"a876e277-510e-4d8a-9c59-a7517de82dba\") " pod="openshift-infra/auto-csr-approver-29551460-4qpgm" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.349755 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzvd6\" (UniqueName: \"kubernetes.io/projected/a876e277-510e-4d8a-9c59-a7517de82dba-kube-api-access-wzvd6\") pod \"auto-csr-approver-29551460-4qpgm\" (UID: \"a876e277-510e-4d8a-9c59-a7517de82dba\") " pod="openshift-infra/auto-csr-approver-29551460-4qpgm" Mar 09 20:20:00 crc kubenswrapper[4750]: I0309 20:20:00.512568 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551460-4qpgm" Mar 09 20:20:01 crc kubenswrapper[4750]: I0309 20:20:01.024819 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551460-4qpgm"] Mar 09 20:20:01 crc kubenswrapper[4750]: I0309 20:20:01.660949 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551460-4qpgm" event={"ID":"a876e277-510e-4d8a-9c59-a7517de82dba","Type":"ContainerStarted","Data":"7f876390b834c3c1322d071660c4de73645d1e29ffc8a83ddbf55aaea76a6bb2"} Mar 09 20:20:02 crc kubenswrapper[4750]: I0309 20:20:02.373518 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:20:02 crc kubenswrapper[4750]: E0309 20:20:02.374061 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:20:02 crc kubenswrapper[4750]: I0309 20:20:02.669718 4750 generic.go:334] "Generic (PLEG): container finished" podID="a876e277-510e-4d8a-9c59-a7517de82dba" containerID="02aa286c93e85be0d2f6423a8393a00834963ee8ffb975e6d59c50ddeef08055" exitCode=0 Mar 09 20:20:02 crc kubenswrapper[4750]: I0309 20:20:02.669833 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551460-4qpgm" event={"ID":"a876e277-510e-4d8a-9c59-a7517de82dba","Type":"ContainerDied","Data":"02aa286c93e85be0d2f6423a8393a00834963ee8ffb975e6d59c50ddeef08055"} Mar 09 20:20:04 crc kubenswrapper[4750]: I0309 20:20:04.131919 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551460-4qpgm" Mar 09 20:20:04 crc kubenswrapper[4750]: I0309 20:20:04.326265 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzvd6\" (UniqueName: \"kubernetes.io/projected/a876e277-510e-4d8a-9c59-a7517de82dba-kube-api-access-wzvd6\") pod \"a876e277-510e-4d8a-9c59-a7517de82dba\" (UID: \"a876e277-510e-4d8a-9c59-a7517de82dba\") " Mar 09 20:20:04 crc kubenswrapper[4750]: I0309 20:20:04.332782 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a876e277-510e-4d8a-9c59-a7517de82dba-kube-api-access-wzvd6" (OuterVolumeSpecName: "kube-api-access-wzvd6") pod "a876e277-510e-4d8a-9c59-a7517de82dba" (UID: "a876e277-510e-4d8a-9c59-a7517de82dba"). InnerVolumeSpecName "kube-api-access-wzvd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:20:04 crc kubenswrapper[4750]: I0309 20:20:04.430128 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzvd6\" (UniqueName: \"kubernetes.io/projected/a876e277-510e-4d8a-9c59-a7517de82dba-kube-api-access-wzvd6\") on node \"crc\" DevicePath \"\"" Mar 09 20:20:04 crc kubenswrapper[4750]: I0309 20:20:04.695347 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551460-4qpgm" event={"ID":"a876e277-510e-4d8a-9c59-a7517de82dba","Type":"ContainerDied","Data":"7f876390b834c3c1322d071660c4de73645d1e29ffc8a83ddbf55aaea76a6bb2"} Mar 09 20:20:04 crc kubenswrapper[4750]: I0309 20:20:04.695415 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f876390b834c3c1322d071660c4de73645d1e29ffc8a83ddbf55aaea76a6bb2" Mar 09 20:20:04 crc kubenswrapper[4750]: I0309 20:20:04.695429 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551460-4qpgm" Mar 09 20:20:05 crc kubenswrapper[4750]: I0309 20:20:05.202810 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551454-7vvml"] Mar 09 20:20:05 crc kubenswrapper[4750]: I0309 20:20:05.213817 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551454-7vvml"] Mar 09 20:20:05 crc kubenswrapper[4750]: I0309 20:20:05.389663 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf60d44-00ad-45fb-b258-c309af142f3b" path="/var/lib/kubelet/pods/cdf60d44-00ad-45fb-b258-c309af142f3b/volumes" Mar 09 20:20:07 crc kubenswrapper[4750]: I0309 20:20:07.373832 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:20:07 crc kubenswrapper[4750]: E0309 20:20:07.374359 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:20:15 crc kubenswrapper[4750]: I0309 20:20:15.374203 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:20:15 crc kubenswrapper[4750]: E0309 20:20:15.375672 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:20:22 crc kubenswrapper[4750]: I0309 20:20:22.374205 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:20:22 crc kubenswrapper[4750]: E0309 20:20:22.375139 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:20:29 crc kubenswrapper[4750]: I0309 20:20:29.382708 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:20:29 crc kubenswrapper[4750]: E0309 20:20:29.384035 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:20:35 crc kubenswrapper[4750]: I0309 20:20:35.377151 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:20:35 crc kubenswrapper[4750]: E0309 20:20:35.379973 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:20:40 crc kubenswrapper[4750]: I0309 20:20:40.373202 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:20:40 crc kubenswrapper[4750]: E0309 20:20:40.374025 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:20:49 crc kubenswrapper[4750]: I0309 20:20:49.385552 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:20:49 crc kubenswrapper[4750]: E0309 20:20:49.386233 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:20:52 crc kubenswrapper[4750]: I0309 20:20:52.373793 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:20:52 crc kubenswrapper[4750]: E0309 20:20:52.375095 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:20:53 crc kubenswrapper[4750]: I0309 20:20:53.326318 4750 scope.go:117] "RemoveContainer" containerID="fa36304e4581e5b87c02d4446a9a18331d63acaee77d8fdbc266ebd9112f4ee5" Mar 09 20:21:04 crc kubenswrapper[4750]: I0309 20:21:04.373980 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:21:04 crc kubenswrapper[4750]: E0309 20:21:04.374606 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:21:05 crc kubenswrapper[4750]: I0309 20:21:05.373739 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:21:05 crc kubenswrapper[4750]: E0309 20:21:05.374587 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:21:15 crc kubenswrapper[4750]: I0309 20:21:15.374425 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:21:15 crc kubenswrapper[4750]: E0309 20:21:15.375702 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:21:16 crc kubenswrapper[4750]: I0309 20:21:16.373856 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:21:16 crc kubenswrapper[4750]: E0309 20:21:16.374750 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:21:27 crc kubenswrapper[4750]: I0309 20:21:27.378195 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:21:27 crc kubenswrapper[4750]: E0309 20:21:27.379379 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:21:30 crc kubenswrapper[4750]: I0309 20:21:30.373661 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:21:30 crc kubenswrapper[4750]: E0309 20:21:30.374475 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:21:42 crc kubenswrapper[4750]: I0309 20:21:42.375763 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:21:42 crc kubenswrapper[4750]: E0309 20:21:42.376987 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:21:45 crc kubenswrapper[4750]: I0309 20:21:45.373822 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:21:46 crc kubenswrapper[4750]: I0309 20:21:46.897333 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2"} Mar 09 20:21:48 crc kubenswrapper[4750]: I0309 20:21:48.930846 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" exitCode=1 Mar 09 20:21:48 crc kubenswrapper[4750]: I0309 20:21:48.930917 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2"} Mar 09 20:21:48 crc kubenswrapper[4750]: I0309 20:21:48.931466 4750 scope.go:117] "RemoveContainer" containerID="0d44f8fd2a898d2278debf8207dc1f5f6ae884a3edb0444d65042e8b1a070400" Mar 09 20:21:48 crc kubenswrapper[4750]: I0309 20:21:48.932949 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:21:48 crc kubenswrapper[4750]: E0309 20:21:48.933685 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:21:50 crc kubenswrapper[4750]: I0309 20:21:50.189907 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:21:50 crc kubenswrapper[4750]: I0309 20:21:50.191127 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:21:50 crc kubenswrapper[4750]: E0309 20:21:50.191494 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:21:50 crc kubenswrapper[4750]: I0309 20:21:50.191757 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:21:50 crc kubenswrapper[4750]: I0309 20:21:50.191805 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:21:50 crc kubenswrapper[4750]: I0309 20:21:50.972975 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:21:50 crc kubenswrapper[4750]: E0309 20:21:50.973796 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:21:53 crc kubenswrapper[4750]: I0309 20:21:53.374616 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:21:53 crc kubenswrapper[4750]: E0309 20:21:53.375594 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:21:59 crc kubenswrapper[4750]: I0309 20:21:59.074037 4750 generic.go:334] "Generic (PLEG): container finished" podID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerID="8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99" exitCode=0 Mar 09 20:21:59 crc kubenswrapper[4750]: I0309 20:21:59.074173 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" event={"ID":"07185cb9-e46d-4abc-9ad2-55b113c0a29f","Type":"ContainerDied","Data":"8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99"} Mar 09 20:21:59 crc kubenswrapper[4750]: I0309 20:21:59.075531 4750 scope.go:117] "RemoveContainer" containerID="8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99" Mar 09 20:21:59 crc kubenswrapper[4750]: I0309 20:21:59.801530 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pkxq8_must-gather-wnhhg_07185cb9-e46d-4abc-9ad2-55b113c0a29f/gather/0.log" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.153109 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551462-9xp8h"] Mar 09 20:22:00 crc kubenswrapper[4750]: E0309 20:22:00.153993 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a876e277-510e-4d8a-9c59-a7517de82dba" containerName="oc" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.154013 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a876e277-510e-4d8a-9c59-a7517de82dba" containerName="oc" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.154226 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a876e277-510e-4d8a-9c59-a7517de82dba" containerName="oc" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.154977 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551462-9xp8h" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.157037 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.157249 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.162837 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.165881 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551462-9xp8h"] Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.252409 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24n8c\" (UniqueName: \"kubernetes.io/projected/43c7203f-3d34-4fc1-970a-18f00646df88-kube-api-access-24n8c\") pod \"auto-csr-approver-29551462-9xp8h\" (UID: \"43c7203f-3d34-4fc1-970a-18f00646df88\") " pod="openshift-infra/auto-csr-approver-29551462-9xp8h" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.356130 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24n8c\" (UniqueName: \"kubernetes.io/projected/43c7203f-3d34-4fc1-970a-18f00646df88-kube-api-access-24n8c\") pod \"auto-csr-approver-29551462-9xp8h\" (UID: \"43c7203f-3d34-4fc1-970a-18f00646df88\") " pod="openshift-infra/auto-csr-approver-29551462-9xp8h" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.373132 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24n8c\" (UniqueName: \"kubernetes.io/projected/43c7203f-3d34-4fc1-970a-18f00646df88-kube-api-access-24n8c\") pod \"auto-csr-approver-29551462-9xp8h\" (UID: \"43c7203f-3d34-4fc1-970a-18f00646df88\") " pod="openshift-infra/auto-csr-approver-29551462-9xp8h" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.479670 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551462-9xp8h" Mar 09 20:22:00 crc kubenswrapper[4750]: I0309 20:22:00.981483 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551462-9xp8h"] Mar 09 20:22:01 crc kubenswrapper[4750]: I0309 20:22:01.097350 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551462-9xp8h" event={"ID":"43c7203f-3d34-4fc1-970a-18f00646df88","Type":"ContainerStarted","Data":"940fe3cfa56c38db8df2198627eb3d88af82de0f8f407409f533a497ffbf1d7e"} Mar 09 20:22:03 crc kubenswrapper[4750]: I0309 20:22:03.125738 4750 generic.go:334] "Generic (PLEG): container finished" podID="43c7203f-3d34-4fc1-970a-18f00646df88" containerID="794754c8ad47d46ca85f09125978a3e0633c56a7a36682a87ea2b57583a7b26f" exitCode=0 Mar 09 20:22:03 crc kubenswrapper[4750]: I0309 20:22:03.125809 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551462-9xp8h" event={"ID":"43c7203f-3d34-4fc1-970a-18f00646df88","Type":"ContainerDied","Data":"794754c8ad47d46ca85f09125978a3e0633c56a7a36682a87ea2b57583a7b26f"} Mar 09 20:22:04 crc kubenswrapper[4750]: I0309 20:22:04.516889 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551462-9xp8h" Mar 09 20:22:04 crc kubenswrapper[4750]: I0309 20:22:04.676390 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24n8c\" (UniqueName: \"kubernetes.io/projected/43c7203f-3d34-4fc1-970a-18f00646df88-kube-api-access-24n8c\") pod \"43c7203f-3d34-4fc1-970a-18f00646df88\" (UID: \"43c7203f-3d34-4fc1-970a-18f00646df88\") " Mar 09 20:22:04 crc kubenswrapper[4750]: I0309 20:22:04.691981 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c7203f-3d34-4fc1-970a-18f00646df88-kube-api-access-24n8c" (OuterVolumeSpecName: "kube-api-access-24n8c") pod "43c7203f-3d34-4fc1-970a-18f00646df88" (UID: "43c7203f-3d34-4fc1-970a-18f00646df88"). InnerVolumeSpecName "kube-api-access-24n8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:22:04 crc kubenswrapper[4750]: I0309 20:22:04.779786 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24n8c\" (UniqueName: \"kubernetes.io/projected/43c7203f-3d34-4fc1-970a-18f00646df88-kube-api-access-24n8c\") on node \"crc\" DevicePath \"\"" Mar 09 20:22:05 crc kubenswrapper[4750]: I0309 20:22:05.161073 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551462-9xp8h" event={"ID":"43c7203f-3d34-4fc1-970a-18f00646df88","Type":"ContainerDied","Data":"940fe3cfa56c38db8df2198627eb3d88af82de0f8f407409f533a497ffbf1d7e"} Mar 09 20:22:05 crc kubenswrapper[4750]: I0309 20:22:05.161136 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="940fe3cfa56c38db8df2198627eb3d88af82de0f8f407409f533a497ffbf1d7e" Mar 09 20:22:05 crc kubenswrapper[4750]: I0309 20:22:05.161129 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551462-9xp8h" Mar 09 20:22:05 crc kubenswrapper[4750]: I0309 20:22:05.376324 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:22:05 crc kubenswrapper[4750]: E0309 20:22:05.380744 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:22:05 crc kubenswrapper[4750]: I0309 20:22:05.588491 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551456-npxpx"] Mar 09 20:22:05 crc kubenswrapper[4750]: I0309 20:22:05.597794 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551456-npxpx"] Mar 09 20:22:07 crc kubenswrapper[4750]: I0309 20:22:07.373253 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:22:07 crc kubenswrapper[4750]: E0309 20:22:07.373569 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:22:07 crc kubenswrapper[4750]: I0309 20:22:07.384494 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="756bf273-1cee-409f-9c91-b45473160359" path="/var/lib/kubelet/pods/756bf273-1cee-409f-9c91-b45473160359/volumes" Mar 09 20:22:07 crc kubenswrapper[4750]: I0309 20:22:07.659012 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pkxq8/must-gather-wnhhg"] Mar 09 20:22:07 crc kubenswrapper[4750]: I0309 20:22:07.659322 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerName="copy" containerID="cri-o://c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb" gracePeriod=2 Mar 09 20:22:07 crc kubenswrapper[4750]: I0309 20:22:07.668351 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pkxq8/must-gather-wnhhg"] Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.101691 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pkxq8_must-gather-wnhhg_07185cb9-e46d-4abc-9ad2-55b113c0a29f/copy/0.log" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.102238 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.190542 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pkxq8_must-gather-wnhhg_07185cb9-e46d-4abc-9ad2-55b113c0a29f/copy/0.log" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.191113 4750 generic.go:334] "Generic (PLEG): container finished" podID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerID="c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb" exitCode=143 Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.191182 4750 scope.go:117] "RemoveContainer" containerID="c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.191347 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pkxq8/must-gather-wnhhg" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.209566 4750 scope.go:117] "RemoveContainer" containerID="8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.246535 4750 scope.go:117] "RemoveContainer" containerID="c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb" Mar 09 20:22:08 crc kubenswrapper[4750]: E0309 20:22:08.247004 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb\": container with ID starting with c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb not found: ID does not exist" containerID="c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.247046 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb"} err="failed to get container status \"c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb\": rpc error: code = NotFound desc = could not find container \"c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb\": container with ID starting with c146c80b946c8d0922cc8ca5ba217794fe014f297a62345d5e8eeb3086990fcb not found: ID does not exist" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.247078 4750 scope.go:117] "RemoveContainer" containerID="8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99" Mar 09 20:22:08 crc kubenswrapper[4750]: E0309 20:22:08.247372 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99\": container with ID starting with 8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99 not found: ID does not exist" containerID="8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.247399 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99"} err="failed to get container status \"8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99\": rpc error: code = NotFound desc = could not find container \"8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99\": container with ID starting with 8da54e6fc9a3b41d65501ad290794b6efb2b5af745ae8754e572d3aee844dd99 not found: ID does not exist" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.252658 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xvnx\" (UniqueName: \"kubernetes.io/projected/07185cb9-e46d-4abc-9ad2-55b113c0a29f-kube-api-access-9xvnx\") pod \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.252895 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/07185cb9-e46d-4abc-9ad2-55b113c0a29f-must-gather-output\") pod \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\" (UID: \"07185cb9-e46d-4abc-9ad2-55b113c0a29f\") " Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.259775 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07185cb9-e46d-4abc-9ad2-55b113c0a29f-kube-api-access-9xvnx" (OuterVolumeSpecName: "kube-api-access-9xvnx") pod "07185cb9-e46d-4abc-9ad2-55b113c0a29f" (UID: "07185cb9-e46d-4abc-9ad2-55b113c0a29f"). InnerVolumeSpecName "kube-api-access-9xvnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.355162 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xvnx\" (UniqueName: \"kubernetes.io/projected/07185cb9-e46d-4abc-9ad2-55b113c0a29f-kube-api-access-9xvnx\") on node \"crc\" DevicePath \"\"" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.435338 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07185cb9-e46d-4abc-9ad2-55b113c0a29f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "07185cb9-e46d-4abc-9ad2-55b113c0a29f" (UID: "07185cb9-e46d-4abc-9ad2-55b113c0a29f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:22:08 crc kubenswrapper[4750]: I0309 20:22:08.457759 4750 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/07185cb9-e46d-4abc-9ad2-55b113c0a29f-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 09 20:22:09 crc kubenswrapper[4750]: I0309 20:22:09.391227 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" path="/var/lib/kubelet/pods/07185cb9-e46d-4abc-9ad2-55b113c0a29f/volumes" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.024657 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8tp7z"] Mar 09 20:22:12 crc kubenswrapper[4750]: E0309 20:22:12.025420 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c7203f-3d34-4fc1-970a-18f00646df88" containerName="oc" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.025436 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c7203f-3d34-4fc1-970a-18f00646df88" containerName="oc" Mar 09 20:22:12 crc kubenswrapper[4750]: E0309 20:22:12.025453 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerName="copy" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.025461 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerName="copy" Mar 09 20:22:12 crc kubenswrapper[4750]: E0309 20:22:12.025484 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerName="gather" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.025493 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerName="gather" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.025804 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerName="gather" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.025961 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="07185cb9-e46d-4abc-9ad2-55b113c0a29f" containerName="copy" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.025988 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c7203f-3d34-4fc1-970a-18f00646df88" containerName="oc" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.028653 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.044174 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8tp7z"] Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.133005 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpz4j\" (UniqueName: \"kubernetes.io/projected/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-kube-api-access-xpz4j\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.135052 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-catalog-content\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.135239 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-utilities\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.237094 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-catalog-content\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.237227 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-utilities\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.237339 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpz4j\" (UniqueName: \"kubernetes.io/projected/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-kube-api-access-xpz4j\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.238409 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-utilities\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.238874 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-catalog-content\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.257057 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpz4j\" (UniqueName: \"kubernetes.io/projected/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-kube-api-access-xpz4j\") pod \"certified-operators-8tp7z\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.365665 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:12 crc kubenswrapper[4750]: I0309 20:22:12.863186 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8tp7z"] Mar 09 20:22:13 crc kubenswrapper[4750]: I0309 20:22:13.237511 4750 generic.go:334] "Generic (PLEG): container finished" podID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerID="ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5" exitCode=0 Mar 09 20:22:13 crc kubenswrapper[4750]: I0309 20:22:13.237853 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tp7z" event={"ID":"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d","Type":"ContainerDied","Data":"ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5"} Mar 09 20:22:13 crc kubenswrapper[4750]: I0309 20:22:13.237898 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tp7z" event={"ID":"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d","Type":"ContainerStarted","Data":"112d0e1737c9e95d529802bc2e2f473465e4352bcb563b98ad8f33ac3770430e"} Mar 09 20:22:14 crc kubenswrapper[4750]: I0309 20:22:14.248953 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tp7z" event={"ID":"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d","Type":"ContainerStarted","Data":"37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf"} Mar 09 20:22:16 crc kubenswrapper[4750]: I0309 20:22:16.270082 4750 generic.go:334] "Generic (PLEG): container finished" podID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerID="37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf" exitCode=0 Mar 09 20:22:16 crc kubenswrapper[4750]: I0309 20:22:16.270165 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tp7z" event={"ID":"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d","Type":"ContainerDied","Data":"37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf"} Mar 09 20:22:16 crc kubenswrapper[4750]: I0309 20:22:16.373727 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:22:16 crc kubenswrapper[4750]: E0309 20:22:16.373977 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:22:17 crc kubenswrapper[4750]: I0309 20:22:17.283775 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tp7z" event={"ID":"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d","Type":"ContainerStarted","Data":"13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d"} Mar 09 20:22:17 crc kubenswrapper[4750]: I0309 20:22:17.307246 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8tp7z" podStartSLOduration=2.910345536 podStartE2EDuration="6.307233551s" podCreationTimestamp="2026-03-09 20:22:11 +0000 UTC" firstStartedPulling="2026-03-09 20:22:13.240058946 +0000 UTC m=+7014.582531344" lastFinishedPulling="2026-03-09 20:22:16.636946961 +0000 UTC m=+7017.979419359" observedRunningTime="2026-03-09 20:22:17.302014579 +0000 UTC m=+7018.644486977" watchObservedRunningTime="2026-03-09 20:22:17.307233551 +0000 UTC m=+7018.649705949" Mar 09 20:22:20 crc kubenswrapper[4750]: I0309 20:22:20.373887 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:22:20 crc kubenswrapper[4750]: E0309 20:22:20.374753 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:22:22 crc kubenswrapper[4750]: I0309 20:22:22.366168 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:22 crc kubenswrapper[4750]: I0309 20:22:22.368111 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:22 crc kubenswrapper[4750]: I0309 20:22:22.412815 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:23 crc kubenswrapper[4750]: I0309 20:22:23.405482 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:23 crc kubenswrapper[4750]: I0309 20:22:23.469405 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8tp7z"] Mar 09 20:22:25 crc kubenswrapper[4750]: I0309 20:22:25.393020 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8tp7z" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="registry-server" containerID="cri-o://13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d" gracePeriod=2 Mar 09 20:22:25 crc kubenswrapper[4750]: I0309 20:22:25.951848 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.125840 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-catalog-content\") pod \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.125967 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpz4j\" (UniqueName: \"kubernetes.io/projected/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-kube-api-access-xpz4j\") pod \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.126015 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-utilities\") pod \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\" (UID: \"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d\") " Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.128545 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-utilities" (OuterVolumeSpecName: "utilities") pod "abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" (UID: "abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.140020 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-kube-api-access-xpz4j" (OuterVolumeSpecName: "kube-api-access-xpz4j") pod "abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" (UID: "abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d"). InnerVolumeSpecName "kube-api-access-xpz4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.203059 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" (UID: "abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.228752 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.228789 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpz4j\" (UniqueName: \"kubernetes.io/projected/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-kube-api-access-xpz4j\") on node \"crc\" DevicePath \"\"" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.228799 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.435549 4750 generic.go:334] "Generic (PLEG): container finished" podID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerID="13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d" exitCode=0 Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.435600 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tp7z" event={"ID":"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d","Type":"ContainerDied","Data":"13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d"} Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.435643 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tp7z" event={"ID":"abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d","Type":"ContainerDied","Data":"112d0e1737c9e95d529802bc2e2f473465e4352bcb563b98ad8f33ac3770430e"} Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.435663 4750 scope.go:117] "RemoveContainer" containerID="13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.435831 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tp7z" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.454388 4750 scope.go:117] "RemoveContainer" containerID="37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.486528 4750 scope.go:117] "RemoveContainer" containerID="ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.495217 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8tp7z"] Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.507211 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8tp7z"] Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.536848 4750 scope.go:117] "RemoveContainer" containerID="13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d" Mar 09 20:22:26 crc kubenswrapper[4750]: E0309 20:22:26.537427 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d\": container with ID starting with 13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d not found: ID does not exist" containerID="13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.537487 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d"} err="failed to get container status \"13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d\": rpc error: code = NotFound desc = could not find container \"13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d\": container with ID starting with 13c0bc5cac42e80d94143e8dc450e50e1f4c0edf28f0aaf7e4d7ab3236da725d not found: ID does not exist" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.537516 4750 scope.go:117] "RemoveContainer" containerID="37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf" Mar 09 20:22:26 crc kubenswrapper[4750]: E0309 20:22:26.537980 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf\": container with ID starting with 37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf not found: ID does not exist" containerID="37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.538015 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf"} err="failed to get container status \"37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf\": rpc error: code = NotFound desc = could not find container \"37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf\": container with ID starting with 37c15fce308e2ff78405725aac85c859442d53e4c33bf50f2d044c4d10ca6fcf not found: ID does not exist" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.538039 4750 scope.go:117] "RemoveContainer" containerID="ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5" Mar 09 20:22:26 crc kubenswrapper[4750]: E0309 20:22:26.538466 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5\": container with ID starting with ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5 not found: ID does not exist" containerID="ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5" Mar 09 20:22:26 crc kubenswrapper[4750]: I0309 20:22:26.538494 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5"} err="failed to get container status \"ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5\": rpc error: code = NotFound desc = could not find container \"ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5\": container with ID starting with ccd224714f727dc1f95d90bcd7a15e1b8841750a9e46dd89ccc95e459048dad5 not found: ID does not exist" Mar 09 20:22:27 crc kubenswrapper[4750]: I0309 20:22:27.383451 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" path="/var/lib/kubelet/pods/abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d/volumes" Mar 09 20:22:28 crc kubenswrapper[4750]: I0309 20:22:28.373497 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:22:28 crc kubenswrapper[4750]: E0309 20:22:28.374048 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:22:35 crc kubenswrapper[4750]: I0309 20:22:35.385478 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:22:35 crc kubenswrapper[4750]: E0309 20:22:35.387151 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:22:40 crc kubenswrapper[4750]: I0309 20:22:40.373650 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:22:40 crc kubenswrapper[4750]: E0309 20:22:40.374732 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:22:48 crc kubenswrapper[4750]: I0309 20:22:48.373224 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:22:48 crc kubenswrapper[4750]: E0309 20:22:48.373995 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:22:53 crc kubenswrapper[4750]: I0309 20:22:53.451170 4750 scope.go:117] "RemoveContainer" containerID="dc23350786883987400b43a67df16f43ed991845ec0e61903a312c09dd49c83e" Mar 09 20:22:55 crc kubenswrapper[4750]: I0309 20:22:55.374336 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:22:55 crc kubenswrapper[4750]: E0309 20:22:55.374805 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.188487 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rqldd"] Mar 09 20:22:59 crc kubenswrapper[4750]: E0309 20:22:59.190068 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="extract-utilities" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.190193 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="extract-utilities" Mar 09 20:22:59 crc kubenswrapper[4750]: E0309 20:22:59.190307 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="extract-content" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.190393 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="extract-content" Mar 09 20:22:59 crc kubenswrapper[4750]: E0309 20:22:59.190548 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="registry-server" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.190653 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="registry-server" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.191025 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="abed56d0-f30c-4a7c-b158-e4fcc2f5ce6d" containerName="registry-server" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.192908 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.201241 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqldd"] Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.290353 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-catalog-content\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.290408 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dqz8\" (UniqueName: \"kubernetes.io/projected/d3dc276e-bded-4911-ae38-acb18a764306-kube-api-access-5dqz8\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.290538 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-utilities\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.392777 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-catalog-content\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.393311 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dqz8\" (UniqueName: \"kubernetes.io/projected/d3dc276e-bded-4911-ae38-acb18a764306-kube-api-access-5dqz8\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.393562 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-utilities\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.394158 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-utilities\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.394451 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-catalog-content\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.417358 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dqz8\" (UniqueName: \"kubernetes.io/projected/d3dc276e-bded-4911-ae38-acb18a764306-kube-api-access-5dqz8\") pod \"community-operators-rqldd\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:22:59 crc kubenswrapper[4750]: I0309 20:22:59.514112 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:23:00 crc kubenswrapper[4750]: I0309 20:23:00.127742 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqldd"] Mar 09 20:23:00 crc kubenswrapper[4750]: I0309 20:23:00.374125 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:23:00 crc kubenswrapper[4750]: E0309 20:23:00.374730 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:23:00 crc kubenswrapper[4750]: I0309 20:23:00.760970 4750 generic.go:334] "Generic (PLEG): container finished" podID="d3dc276e-bded-4911-ae38-acb18a764306" containerID="55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4" exitCode=0 Mar 09 20:23:00 crc kubenswrapper[4750]: I0309 20:23:00.761029 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqldd" event={"ID":"d3dc276e-bded-4911-ae38-acb18a764306","Type":"ContainerDied","Data":"55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4"} Mar 09 20:23:00 crc kubenswrapper[4750]: I0309 20:23:00.761354 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqldd" event={"ID":"d3dc276e-bded-4911-ae38-acb18a764306","Type":"ContainerStarted","Data":"a2b1621777f3e224c2af10cf28047fd585fcec6bca313a757ee0da86e8499a43"} Mar 09 20:23:01 crc kubenswrapper[4750]: I0309 20:23:01.772922 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqldd" event={"ID":"d3dc276e-bded-4911-ae38-acb18a764306","Type":"ContainerStarted","Data":"6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312"} Mar 09 20:23:03 crc kubenswrapper[4750]: I0309 20:23:03.791540 4750 generic.go:334] "Generic (PLEG): container finished" podID="d3dc276e-bded-4911-ae38-acb18a764306" containerID="6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312" exitCode=0 Mar 09 20:23:03 crc kubenswrapper[4750]: I0309 20:23:03.791641 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqldd" event={"ID":"d3dc276e-bded-4911-ae38-acb18a764306","Type":"ContainerDied","Data":"6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312"} Mar 09 20:23:03 crc kubenswrapper[4750]: I0309 20:23:03.794080 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 20:23:04 crc kubenswrapper[4750]: I0309 20:23:04.804161 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqldd" event={"ID":"d3dc276e-bded-4911-ae38-acb18a764306","Type":"ContainerStarted","Data":"da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8"} Mar 09 20:23:04 crc kubenswrapper[4750]: I0309 20:23:04.830944 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rqldd" podStartSLOduration=2.392369776 podStartE2EDuration="5.830926738s" podCreationTimestamp="2026-03-09 20:22:59 +0000 UTC" firstStartedPulling="2026-03-09 20:23:00.762961913 +0000 UTC m=+7062.105434311" lastFinishedPulling="2026-03-09 20:23:04.201518875 +0000 UTC m=+7065.543991273" observedRunningTime="2026-03-09 20:23:04.821258306 +0000 UTC m=+7066.163730714" watchObservedRunningTime="2026-03-09 20:23:04.830926738 +0000 UTC m=+7066.173399136" Mar 09 20:23:06 crc kubenswrapper[4750]: I0309 20:23:06.374194 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:23:06 crc kubenswrapper[4750]: E0309 20:23:06.374438 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:23:09 crc kubenswrapper[4750]: I0309 20:23:09.514582 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:23:09 crc kubenswrapper[4750]: I0309 20:23:09.514889 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:23:09 crc kubenswrapper[4750]: I0309 20:23:09.563505 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:23:09 crc kubenswrapper[4750]: I0309 20:23:09.889414 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:23:09 crc kubenswrapper[4750]: I0309 20:23:09.941060 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqldd"] Mar 09 20:23:11 crc kubenswrapper[4750]: I0309 20:23:11.872122 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rqldd" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="registry-server" containerID="cri-o://da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8" gracePeriod=2 Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.388014 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.466363 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-utilities\") pod \"d3dc276e-bded-4911-ae38-acb18a764306\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.466534 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-catalog-content\") pod \"d3dc276e-bded-4911-ae38-acb18a764306\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.466697 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dqz8\" (UniqueName: \"kubernetes.io/projected/d3dc276e-bded-4911-ae38-acb18a764306-kube-api-access-5dqz8\") pod \"d3dc276e-bded-4911-ae38-acb18a764306\" (UID: \"d3dc276e-bded-4911-ae38-acb18a764306\") " Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.468338 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-utilities" (OuterVolumeSpecName: "utilities") pod "d3dc276e-bded-4911-ae38-acb18a764306" (UID: "d3dc276e-bded-4911-ae38-acb18a764306"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.468662 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.476532 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3dc276e-bded-4911-ae38-acb18a764306-kube-api-access-5dqz8" (OuterVolumeSpecName: "kube-api-access-5dqz8") pod "d3dc276e-bded-4911-ae38-acb18a764306" (UID: "d3dc276e-bded-4911-ae38-acb18a764306"). InnerVolumeSpecName "kube-api-access-5dqz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.526744 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3dc276e-bded-4911-ae38-acb18a764306" (UID: "d3dc276e-bded-4911-ae38-acb18a764306"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.570114 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3dc276e-bded-4911-ae38-acb18a764306-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.570385 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dqz8\" (UniqueName: \"kubernetes.io/projected/d3dc276e-bded-4911-ae38-acb18a764306-kube-api-access-5dqz8\") on node \"crc\" DevicePath \"\"" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.886840 4750 generic.go:334] "Generic (PLEG): container finished" podID="d3dc276e-bded-4911-ae38-acb18a764306" containerID="da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8" exitCode=0 Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.886899 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqldd" event={"ID":"d3dc276e-bded-4911-ae38-acb18a764306","Type":"ContainerDied","Data":"da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8"} Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.886934 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqldd" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.886971 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqldd" event={"ID":"d3dc276e-bded-4911-ae38-acb18a764306","Type":"ContainerDied","Data":"a2b1621777f3e224c2af10cf28047fd585fcec6bca313a757ee0da86e8499a43"} Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.887000 4750 scope.go:117] "RemoveContainer" containerID="da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.910400 4750 scope.go:117] "RemoveContainer" containerID="6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.935610 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqldd"] Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.956905 4750 scope.go:117] "RemoveContainer" containerID="55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.957149 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rqldd"] Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.992536 4750 scope.go:117] "RemoveContainer" containerID="da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8" Mar 09 20:23:12 crc kubenswrapper[4750]: E0309 20:23:12.992918 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8\": container with ID starting with da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8 not found: ID does not exist" containerID="da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.992957 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8"} err="failed to get container status \"da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8\": rpc error: code = NotFound desc = could not find container \"da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8\": container with ID starting with da87f1c5dee3f8eebb65c33e6bf3ce782f45729129f39afc7b5b40d0ef02f4b8 not found: ID does not exist" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.992981 4750 scope.go:117] "RemoveContainer" containerID="6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312" Mar 09 20:23:12 crc kubenswrapper[4750]: E0309 20:23:12.993153 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312\": container with ID starting with 6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312 not found: ID does not exist" containerID="6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.993172 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312"} err="failed to get container status \"6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312\": rpc error: code = NotFound desc = could not find container \"6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312\": container with ID starting with 6bbbad14ba792af6547788cedcc7eab70d70750b484e38781ea2407a60974312 not found: ID does not exist" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.993187 4750 scope.go:117] "RemoveContainer" containerID="55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4" Mar 09 20:23:12 crc kubenswrapper[4750]: E0309 20:23:12.993332 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4\": container with ID starting with 55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4 not found: ID does not exist" containerID="55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4" Mar 09 20:23:12 crc kubenswrapper[4750]: I0309 20:23:12.993350 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4"} err="failed to get container status \"55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4\": rpc error: code = NotFound desc = could not find container \"55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4\": container with ID starting with 55ac3226b7094701388a687dd38743227715aa7d73a0c145b3ef4d3bbf7e3aa4 not found: ID does not exist" Mar 09 20:23:13 crc kubenswrapper[4750]: I0309 20:23:13.373618 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:23:13 crc kubenswrapper[4750]: E0309 20:23:13.373929 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pqlpj_openshift-machine-config-operator(c7f6ff1a-bc64-466a-9f65-59acfede7fc1)\"" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" Mar 09 20:23:13 crc kubenswrapper[4750]: I0309 20:23:13.385270 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3dc276e-bded-4911-ae38-acb18a764306" path="/var/lib/kubelet/pods/d3dc276e-bded-4911-ae38-acb18a764306/volumes" Mar 09 20:23:19 crc kubenswrapper[4750]: I0309 20:23:19.386573 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:23:19 crc kubenswrapper[4750]: E0309 20:23:19.387256 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:23:27 crc kubenswrapper[4750]: I0309 20:23:27.372865 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:23:28 crc kubenswrapper[4750]: I0309 20:23:28.045555 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"4b588e8af4c473b918ca911bdd54d8c2c3ec5f320166fb8ba7cb0a5d48585a5f"} Mar 09 20:23:31 crc kubenswrapper[4750]: I0309 20:23:31.374256 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:23:31 crc kubenswrapper[4750]: E0309 20:23:31.375096 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.130122 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w8w2d"] Mar 09 20:23:33 crc kubenswrapper[4750]: E0309 20:23:33.132196 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="extract-content" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.132224 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="extract-content" Mar 09 20:23:33 crc kubenswrapper[4750]: E0309 20:23:33.132248 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="extract-utilities" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.132266 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="extract-utilities" Mar 09 20:23:33 crc kubenswrapper[4750]: E0309 20:23:33.132296 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="registry-server" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.132304 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="registry-server" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.132576 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3dc276e-bded-4911-ae38-acb18a764306" containerName="registry-server" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.134661 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.148672 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8w2d"] Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.240607 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-utilities\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.240885 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqw7h\" (UniqueName: \"kubernetes.io/projected/4630e330-67f7-4b53-a02a-6c6d85163888-kube-api-access-kqw7h\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.241076 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-catalog-content\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.343027 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-utilities\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.343069 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqw7h\" (UniqueName: \"kubernetes.io/projected/4630e330-67f7-4b53-a02a-6c6d85163888-kube-api-access-kqw7h\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.343173 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-catalog-content\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.343564 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-utilities\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.343645 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-catalog-content\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.365038 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqw7h\" (UniqueName: \"kubernetes.io/projected/4630e330-67f7-4b53-a02a-6c6d85163888-kube-api-access-kqw7h\") pod \"redhat-operators-w8w2d\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.479406 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:33 crc kubenswrapper[4750]: I0309 20:23:33.969645 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8w2d"] Mar 09 20:23:34 crc kubenswrapper[4750]: I0309 20:23:34.120261 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8w2d" event={"ID":"4630e330-67f7-4b53-a02a-6c6d85163888","Type":"ContainerStarted","Data":"d1903c99b416db43e1d77cb7638c00cc0f08334dc74a39d6ad0e108aa3082e3d"} Mar 09 20:23:35 crc kubenswrapper[4750]: I0309 20:23:35.135331 4750 generic.go:334] "Generic (PLEG): container finished" podID="4630e330-67f7-4b53-a02a-6c6d85163888" containerID="622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071" exitCode=0 Mar 09 20:23:35 crc kubenswrapper[4750]: I0309 20:23:35.135383 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8w2d" event={"ID":"4630e330-67f7-4b53-a02a-6c6d85163888","Type":"ContainerDied","Data":"622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071"} Mar 09 20:23:36 crc kubenswrapper[4750]: I0309 20:23:36.147982 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8w2d" event={"ID":"4630e330-67f7-4b53-a02a-6c6d85163888","Type":"ContainerStarted","Data":"97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567"} Mar 09 20:23:42 crc kubenswrapper[4750]: I0309 20:23:42.219664 4750 generic.go:334] "Generic (PLEG): container finished" podID="4630e330-67f7-4b53-a02a-6c6d85163888" containerID="97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567" exitCode=0 Mar 09 20:23:42 crc kubenswrapper[4750]: I0309 20:23:42.219668 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8w2d" event={"ID":"4630e330-67f7-4b53-a02a-6c6d85163888","Type":"ContainerDied","Data":"97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567"} Mar 09 20:23:43 crc kubenswrapper[4750]: I0309 20:23:43.237421 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8w2d" event={"ID":"4630e330-67f7-4b53-a02a-6c6d85163888","Type":"ContainerStarted","Data":"6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3"} Mar 09 20:23:43 crc kubenswrapper[4750]: I0309 20:23:43.272899 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w8w2d" podStartSLOduration=2.789290104 podStartE2EDuration="10.272876942s" podCreationTimestamp="2026-03-09 20:23:33 +0000 UTC" firstStartedPulling="2026-03-09 20:23:35.138598185 +0000 UTC m=+7096.481070583" lastFinishedPulling="2026-03-09 20:23:42.622185003 +0000 UTC m=+7103.964657421" observedRunningTime="2026-03-09 20:23:43.262189853 +0000 UTC m=+7104.604662251" watchObservedRunningTime="2026-03-09 20:23:43.272876942 +0000 UTC m=+7104.615349360" Mar 09 20:23:43 crc kubenswrapper[4750]: I0309 20:23:43.479738 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:43 crc kubenswrapper[4750]: I0309 20:23:43.479798 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:44 crc kubenswrapper[4750]: I0309 20:23:44.533051 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w8w2d" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="registry-server" probeResult="failure" output=< Mar 09 20:23:44 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Mar 09 20:23:44 crc kubenswrapper[4750]: > Mar 09 20:23:45 crc kubenswrapper[4750]: I0309 20:23:45.374686 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:23:45 crc kubenswrapper[4750]: E0309 20:23:45.374942 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:23:53 crc kubenswrapper[4750]: I0309 20:23:53.540242 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:53 crc kubenswrapper[4750]: I0309 20:23:53.603265 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:53 crc kubenswrapper[4750]: I0309 20:23:53.781847 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8w2d"] Mar 09 20:23:55 crc kubenswrapper[4750]: I0309 20:23:55.365863 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w8w2d" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="registry-server" containerID="cri-o://6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3" gracePeriod=2 Mar 09 20:23:55 crc kubenswrapper[4750]: I0309 20:23:55.904039 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.039222 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-catalog-content\") pod \"4630e330-67f7-4b53-a02a-6c6d85163888\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.039407 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-utilities\") pod \"4630e330-67f7-4b53-a02a-6c6d85163888\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.039527 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqw7h\" (UniqueName: \"kubernetes.io/projected/4630e330-67f7-4b53-a02a-6c6d85163888-kube-api-access-kqw7h\") pod \"4630e330-67f7-4b53-a02a-6c6d85163888\" (UID: \"4630e330-67f7-4b53-a02a-6c6d85163888\") " Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.040713 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-utilities" (OuterVolumeSpecName: "utilities") pod "4630e330-67f7-4b53-a02a-6c6d85163888" (UID: "4630e330-67f7-4b53-a02a-6c6d85163888"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.047789 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4630e330-67f7-4b53-a02a-6c6d85163888-kube-api-access-kqw7h" (OuterVolumeSpecName: "kube-api-access-kqw7h") pod "4630e330-67f7-4b53-a02a-6c6d85163888" (UID: "4630e330-67f7-4b53-a02a-6c6d85163888"). InnerVolumeSpecName "kube-api-access-kqw7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.142482 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.142517 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqw7h\" (UniqueName: \"kubernetes.io/projected/4630e330-67f7-4b53-a02a-6c6d85163888-kube-api-access-kqw7h\") on node \"crc\" DevicePath \"\"" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.174630 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4630e330-67f7-4b53-a02a-6c6d85163888" (UID: "4630e330-67f7-4b53-a02a-6c6d85163888"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.244287 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4630e330-67f7-4b53-a02a-6c6d85163888-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.381807 4750 generic.go:334] "Generic (PLEG): container finished" podID="4630e330-67f7-4b53-a02a-6c6d85163888" containerID="6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3" exitCode=0 Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.381904 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8w2d" event={"ID":"4630e330-67f7-4b53-a02a-6c6d85163888","Type":"ContainerDied","Data":"6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3"} Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.382538 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8w2d" event={"ID":"4630e330-67f7-4b53-a02a-6c6d85163888","Type":"ContainerDied","Data":"d1903c99b416db43e1d77cb7638c00cc0f08334dc74a39d6ad0e108aa3082e3d"} Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.381921 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8w2d" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.382569 4750 scope.go:117] "RemoveContainer" containerID="6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.423060 4750 scope.go:117] "RemoveContainer" containerID="97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.428479 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8w2d"] Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.437226 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w8w2d"] Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.452825 4750 scope.go:117] "RemoveContainer" containerID="622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.501167 4750 scope.go:117] "RemoveContainer" containerID="6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3" Mar 09 20:23:56 crc kubenswrapper[4750]: E0309 20:23:56.501923 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3\": container with ID starting with 6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3 not found: ID does not exist" containerID="6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.501981 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3"} err="failed to get container status \"6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3\": rpc error: code = NotFound desc = could not find container \"6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3\": container with ID starting with 6c2b9aa80d8dd57460f47628b0d844e95f85dbbd507d9b6b3b6644ec148db4e3 not found: ID does not exist" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.502021 4750 scope.go:117] "RemoveContainer" containerID="97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567" Mar 09 20:23:56 crc kubenswrapper[4750]: E0309 20:23:56.502397 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567\": container with ID starting with 97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567 not found: ID does not exist" containerID="97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.502438 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567"} err="failed to get container status \"97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567\": rpc error: code = NotFound desc = could not find container \"97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567\": container with ID starting with 97cc48790a84bab9df2fab47f9fd5b595aa36153e6d9390004b175dfac8f4567 not found: ID does not exist" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.502465 4750 scope.go:117] "RemoveContainer" containerID="622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071" Mar 09 20:23:56 crc kubenswrapper[4750]: E0309 20:23:56.502821 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071\": container with ID starting with 622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071 not found: ID does not exist" containerID="622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071" Mar 09 20:23:56 crc kubenswrapper[4750]: I0309 20:23:56.502885 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071"} err="failed to get container status \"622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071\": rpc error: code = NotFound desc = could not find container \"622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071\": container with ID starting with 622ec613773adcc7082ea42219ec40dcbd631dcbcb7d6daa0a2eb6bd987ff071 not found: ID does not exist" Mar 09 20:23:57 crc kubenswrapper[4750]: I0309 20:23:57.392401 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" path="/var/lib/kubelet/pods/4630e330-67f7-4b53-a02a-6c6d85163888/volumes" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.168394 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551464-bmpzl"] Mar 09 20:24:00 crc kubenswrapper[4750]: E0309 20:24:00.169124 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="registry-server" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.169141 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="registry-server" Mar 09 20:24:00 crc kubenswrapper[4750]: E0309 20:24:00.169167 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="extract-content" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.169174 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="extract-content" Mar 09 20:24:00 crc kubenswrapper[4750]: E0309 20:24:00.169200 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="extract-utilities" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.169208 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="extract-utilities" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.169603 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4630e330-67f7-4b53-a02a-6c6d85163888" containerName="registry-server" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.171058 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551464-bmpzl" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.176150 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.176249 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.176307 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.183979 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551464-bmpzl"] Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.331362 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnmr2\" (UniqueName: \"kubernetes.io/projected/57b25e8d-20ab-433e-ba9c-c252f83a62b1-kube-api-access-bnmr2\") pod \"auto-csr-approver-29551464-bmpzl\" (UID: \"57b25e8d-20ab-433e-ba9c-c252f83a62b1\") " pod="openshift-infra/auto-csr-approver-29551464-bmpzl" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.373453 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:24:00 crc kubenswrapper[4750]: E0309 20:24:00.373840 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.433107 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnmr2\" (UniqueName: \"kubernetes.io/projected/57b25e8d-20ab-433e-ba9c-c252f83a62b1-kube-api-access-bnmr2\") pod \"auto-csr-approver-29551464-bmpzl\" (UID: \"57b25e8d-20ab-433e-ba9c-c252f83a62b1\") " pod="openshift-infra/auto-csr-approver-29551464-bmpzl" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.464593 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnmr2\" (UniqueName: \"kubernetes.io/projected/57b25e8d-20ab-433e-ba9c-c252f83a62b1-kube-api-access-bnmr2\") pod \"auto-csr-approver-29551464-bmpzl\" (UID: \"57b25e8d-20ab-433e-ba9c-c252f83a62b1\") " pod="openshift-infra/auto-csr-approver-29551464-bmpzl" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.514618 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551464-bmpzl" Mar 09 20:24:00 crc kubenswrapper[4750]: I0309 20:24:00.955805 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551464-bmpzl"] Mar 09 20:24:01 crc kubenswrapper[4750]: I0309 20:24:01.080230 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551464-bmpzl" event={"ID":"57b25e8d-20ab-433e-ba9c-c252f83a62b1","Type":"ContainerStarted","Data":"b41620e944ec1c01888fd441241ef8400d7f7ae39b14c82b3edd0b751d06e46a"} Mar 09 20:24:03 crc kubenswrapper[4750]: I0309 20:24:03.103098 4750 generic.go:334] "Generic (PLEG): container finished" podID="57b25e8d-20ab-433e-ba9c-c252f83a62b1" containerID="d4372d754b8ab56eeee6f2224150e078e3c28564a60851bafd0798a05a8a512d" exitCode=0 Mar 09 20:24:03 crc kubenswrapper[4750]: I0309 20:24:03.103153 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551464-bmpzl" event={"ID":"57b25e8d-20ab-433e-ba9c-c252f83a62b1","Type":"ContainerDied","Data":"d4372d754b8ab56eeee6f2224150e078e3c28564a60851bafd0798a05a8a512d"} Mar 09 20:24:04 crc kubenswrapper[4750]: I0309 20:24:04.495493 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551464-bmpzl" Mar 09 20:24:04 crc kubenswrapper[4750]: I0309 20:24:04.616115 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnmr2\" (UniqueName: \"kubernetes.io/projected/57b25e8d-20ab-433e-ba9c-c252f83a62b1-kube-api-access-bnmr2\") pod \"57b25e8d-20ab-433e-ba9c-c252f83a62b1\" (UID: \"57b25e8d-20ab-433e-ba9c-c252f83a62b1\") " Mar 09 20:24:04 crc kubenswrapper[4750]: I0309 20:24:04.628006 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b25e8d-20ab-433e-ba9c-c252f83a62b1-kube-api-access-bnmr2" (OuterVolumeSpecName: "kube-api-access-bnmr2") pod "57b25e8d-20ab-433e-ba9c-c252f83a62b1" (UID: "57b25e8d-20ab-433e-ba9c-c252f83a62b1"). InnerVolumeSpecName "kube-api-access-bnmr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:24:04 crc kubenswrapper[4750]: I0309 20:24:04.719226 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnmr2\" (UniqueName: \"kubernetes.io/projected/57b25e8d-20ab-433e-ba9c-c252f83a62b1-kube-api-access-bnmr2\") on node \"crc\" DevicePath \"\"" Mar 09 20:24:05 crc kubenswrapper[4750]: I0309 20:24:05.125660 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551464-bmpzl" event={"ID":"57b25e8d-20ab-433e-ba9c-c252f83a62b1","Type":"ContainerDied","Data":"b41620e944ec1c01888fd441241ef8400d7f7ae39b14c82b3edd0b751d06e46a"} Mar 09 20:24:05 crc kubenswrapper[4750]: I0309 20:24:05.125704 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b41620e944ec1c01888fd441241ef8400d7f7ae39b14c82b3edd0b751d06e46a" Mar 09 20:24:05 crc kubenswrapper[4750]: I0309 20:24:05.125799 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551464-bmpzl" Mar 09 20:24:05 crc kubenswrapper[4750]: I0309 20:24:05.572769 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551458-2pm68"] Mar 09 20:24:05 crc kubenswrapper[4750]: I0309 20:24:05.582392 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551458-2pm68"] Mar 09 20:24:07 crc kubenswrapper[4750]: I0309 20:24:07.386780 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f078a8e-15b4-4a78-8566-0d650ffcbf4c" path="/var/lib/kubelet/pods/6f078a8e-15b4-4a78-8566-0d650ffcbf4c/volumes" Mar 09 20:24:12 crc kubenswrapper[4750]: I0309 20:24:12.374193 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:24:12 crc kubenswrapper[4750]: E0309 20:24:12.375182 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:24:24 crc kubenswrapper[4750]: I0309 20:24:24.373941 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:24:24 crc kubenswrapper[4750]: E0309 20:24:24.374912 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:24:38 crc kubenswrapper[4750]: I0309 20:24:38.373987 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:24:38 crc kubenswrapper[4750]: E0309 20:24:38.374884 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:24:51 crc kubenswrapper[4750]: I0309 20:24:51.374091 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:24:51 crc kubenswrapper[4750]: E0309 20:24:51.375861 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:24:53 crc kubenswrapper[4750]: I0309 20:24:53.638962 4750 scope.go:117] "RemoveContainer" containerID="af2fec232aec7c00cf3be0a2b8c76305b47b2d45b7f62a9249b7bc1793f0946b" Mar 09 20:25:02 crc kubenswrapper[4750]: I0309 20:25:02.374173 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:25:02 crc kubenswrapper[4750]: E0309 20:25:02.375108 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:25:15 crc kubenswrapper[4750]: I0309 20:25:15.373749 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:25:15 crc kubenswrapper[4750]: E0309 20:25:15.374661 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:25:28 crc kubenswrapper[4750]: I0309 20:25:28.373266 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:25:28 crc kubenswrapper[4750]: E0309 20:25:28.375676 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:25:43 crc kubenswrapper[4750]: I0309 20:25:43.373362 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:25:43 crc kubenswrapper[4750]: E0309 20:25:43.374307 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:25:51 crc kubenswrapper[4750]: I0309 20:25:51.744065 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:25:51 crc kubenswrapper[4750]: I0309 20:25:51.744717 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:25:54 crc kubenswrapper[4750]: I0309 20:25:54.414405 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:25:54 crc kubenswrapper[4750]: E0309 20:25:54.415127 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.174935 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29551466-zstlr"] Mar 09 20:26:00 crc kubenswrapper[4750]: E0309 20:26:00.175659 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b25e8d-20ab-433e-ba9c-c252f83a62b1" containerName="oc" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.175674 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b25e8d-20ab-433e-ba9c-c252f83a62b1" containerName="oc" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.175953 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b25e8d-20ab-433e-ba9c-c252f83a62b1" containerName="oc" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.176793 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551466-zstlr" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.180387 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-dftnf" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.180399 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.181014 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.194512 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551466-zstlr"] Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.372793 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqkpk\" (UniqueName: \"kubernetes.io/projected/b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b-kube-api-access-kqkpk\") pod \"auto-csr-approver-29551466-zstlr\" (UID: \"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b\") " pod="openshift-infra/auto-csr-approver-29551466-zstlr" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.477914 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqkpk\" (UniqueName: \"kubernetes.io/projected/b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b-kube-api-access-kqkpk\") pod \"auto-csr-approver-29551466-zstlr\" (UID: \"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b\") " pod="openshift-infra/auto-csr-approver-29551466-zstlr" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.500363 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqkpk\" (UniqueName: \"kubernetes.io/projected/b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b-kube-api-access-kqkpk\") pod \"auto-csr-approver-29551466-zstlr\" (UID: \"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b\") " pod="openshift-infra/auto-csr-approver-29551466-zstlr" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.506537 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551466-zstlr" Mar 09 20:26:00 crc kubenswrapper[4750]: I0309 20:26:00.974499 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29551466-zstlr"] Mar 09 20:26:00 crc kubenswrapper[4750]: W0309 20:26:00.979192 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6ca1b02_2ab6_4e81_bb1b_cad08e055e3b.slice/crio-4cc7f886c08a0257a9c2afea079b5d85b4761c50c2155e153e9781b530716d51 WatchSource:0}: Error finding container 4cc7f886c08a0257a9c2afea079b5d85b4761c50c2155e153e9781b530716d51: Status 404 returned error can't find the container with id 4cc7f886c08a0257a9c2afea079b5d85b4761c50c2155e153e9781b530716d51 Mar 09 20:26:01 crc kubenswrapper[4750]: I0309 20:26:01.498134 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551466-zstlr" event={"ID":"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b","Type":"ContainerStarted","Data":"4cc7f886c08a0257a9c2afea079b5d85b4761c50c2155e153e9781b530716d51"} Mar 09 20:26:02 crc kubenswrapper[4750]: I0309 20:26:02.507013 4750 generic.go:334] "Generic (PLEG): container finished" podID="b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b" containerID="39d9af22cce183988ceb265cdf9483f5b82aefaca548c7f97890521c6c26dcdc" exitCode=0 Mar 09 20:26:02 crc kubenswrapper[4750]: I0309 20:26:02.507056 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551466-zstlr" event={"ID":"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b","Type":"ContainerDied","Data":"39d9af22cce183988ceb265cdf9483f5b82aefaca548c7f97890521c6c26dcdc"} Mar 09 20:26:03 crc kubenswrapper[4750]: I0309 20:26:03.877226 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551466-zstlr" Mar 09 20:26:03 crc kubenswrapper[4750]: I0309 20:26:03.955688 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqkpk\" (UniqueName: \"kubernetes.io/projected/b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b-kube-api-access-kqkpk\") pod \"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b\" (UID: \"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b\") " Mar 09 20:26:03 crc kubenswrapper[4750]: I0309 20:26:03.969337 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b-kube-api-access-kqkpk" (OuterVolumeSpecName: "kube-api-access-kqkpk") pod "b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b" (UID: "b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b"). InnerVolumeSpecName "kube-api-access-kqkpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 20:26:04 crc kubenswrapper[4750]: I0309 20:26:04.059313 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqkpk\" (UniqueName: \"kubernetes.io/projected/b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b-kube-api-access-kqkpk\") on node \"crc\" DevicePath \"\"" Mar 09 20:26:04 crc kubenswrapper[4750]: I0309 20:26:04.525788 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29551466-zstlr" event={"ID":"b6ca1b02-2ab6-4e81-bb1b-cad08e055e3b","Type":"ContainerDied","Data":"4cc7f886c08a0257a9c2afea079b5d85b4761c50c2155e153e9781b530716d51"} Mar 09 20:26:04 crc kubenswrapper[4750]: I0309 20:26:04.525865 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cc7f886c08a0257a9c2afea079b5d85b4761c50c2155e153e9781b530716d51" Mar 09 20:26:04 crc kubenswrapper[4750]: I0309 20:26:04.525835 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29551466-zstlr" Mar 09 20:26:04 crc kubenswrapper[4750]: I0309 20:26:04.973608 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29551460-4qpgm"] Mar 09 20:26:04 crc kubenswrapper[4750]: I0309 20:26:04.988828 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29551460-4qpgm"] Mar 09 20:26:05 crc kubenswrapper[4750]: I0309 20:26:05.374255 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:26:05 crc kubenswrapper[4750]: E0309 20:26:05.374660 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:26:05 crc kubenswrapper[4750]: I0309 20:26:05.388268 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a876e277-510e-4d8a-9c59-a7517de82dba" path="/var/lib/kubelet/pods/a876e277-510e-4d8a-9c59-a7517de82dba/volumes" Mar 09 20:26:16 crc kubenswrapper[4750]: I0309 20:26:16.374190 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:26:16 crc kubenswrapper[4750]: E0309 20:26:16.375686 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:26:21 crc kubenswrapper[4750]: I0309 20:26:21.743868 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:26:21 crc kubenswrapper[4750]: I0309 20:26:21.744302 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:26:28 crc kubenswrapper[4750]: I0309 20:26:28.374396 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:26:28 crc kubenswrapper[4750]: E0309 20:26:28.375515 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:26:40 crc kubenswrapper[4750]: I0309 20:26:40.373516 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:26:40 crc kubenswrapper[4750]: E0309 20:26:40.374504 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:26:51 crc kubenswrapper[4750]: I0309 20:26:51.743425 4750 patch_prober.go:28] interesting pod/machine-config-daemon-pqlpj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 20:26:51 crc kubenswrapper[4750]: I0309 20:26:51.743894 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 20:26:51 crc kubenswrapper[4750]: I0309 20:26:51.743948 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" Mar 09 20:26:51 crc kubenswrapper[4750]: I0309 20:26:51.744812 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b588e8af4c473b918ca911bdd54d8c2c3ec5f320166fb8ba7cb0a5d48585a5f"} pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 20:26:51 crc kubenswrapper[4750]: I0309 20:26:51.744871 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" podUID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerName="machine-config-daemon" containerID="cri-o://4b588e8af4c473b918ca911bdd54d8c2c3ec5f320166fb8ba7cb0a5d48585a5f" gracePeriod=600 Mar 09 20:26:52 crc kubenswrapper[4750]: I0309 20:26:52.041581 4750 generic.go:334] "Generic (PLEG): container finished" podID="c7f6ff1a-bc64-466a-9f65-59acfede7fc1" containerID="4b588e8af4c473b918ca911bdd54d8c2c3ec5f320166fb8ba7cb0a5d48585a5f" exitCode=0 Mar 09 20:26:52 crc kubenswrapper[4750]: I0309 20:26:52.041654 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerDied","Data":"4b588e8af4c473b918ca911bdd54d8c2c3ec5f320166fb8ba7cb0a5d48585a5f"} Mar 09 20:26:52 crc kubenswrapper[4750]: I0309 20:26:52.041933 4750 scope.go:117] "RemoveContainer" containerID="f28fcbe01c60b8e613576e8538a1361dc6f051913c2b5afded4f30b6266d77a8" Mar 09 20:26:53 crc kubenswrapper[4750]: I0309 20:26:53.055098 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pqlpj" event={"ID":"c7f6ff1a-bc64-466a-9f65-59acfede7fc1","Type":"ContainerStarted","Data":"d3c64d9d4d5c71232fcf498b819749c19618bd1abb6c739ae93b7064ebbb9b71"} Mar 09 20:26:53 crc kubenswrapper[4750]: I0309 20:26:53.774930 4750 scope.go:117] "RemoveContainer" containerID="02aa286c93e85be0d2f6423a8393a00834963ee8ffb975e6d59c50ddeef08055" Mar 09 20:26:54 crc kubenswrapper[4750]: I0309 20:26:54.374698 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:26:55 crc kubenswrapper[4750]: I0309 20:26:55.076232 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerStarted","Data":"724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383"} Mar 09 20:26:55 crc kubenswrapper[4750]: I0309 20:26:55.190248 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:26:58 crc kubenswrapper[4750]: I0309 20:26:58.105596 4750 generic.go:334] "Generic (PLEG): container finished" podID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" containerID="724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383" exitCode=1 Mar 09 20:26:58 crc kubenswrapper[4750]: I0309 20:26:58.105815 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2488a7a-65a2-442f-81bf-5c74ce640dc4","Type":"ContainerDied","Data":"724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383"} Mar 09 20:26:58 crc kubenswrapper[4750]: I0309 20:26:58.106427 4750 scope.go:117] "RemoveContainer" containerID="9af2a2ba8657413d91972600e203e78316509f2e6a203d54b6dc6cb5fc4823b2" Mar 09 20:26:58 crc kubenswrapper[4750]: I0309 20:26:58.108761 4750 scope.go:117] "RemoveContainer" containerID="724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383" Mar 09 20:26:58 crc kubenswrapper[4750]: E0309 20:26:58.109886 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:26:59 crc kubenswrapper[4750]: I0309 20:26:59.190359 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:26:59 crc kubenswrapper[4750]: I0309 20:26:59.191514 4750 scope.go:117] "RemoveContainer" containerID="724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383" Mar 09 20:26:59 crc kubenswrapper[4750]: E0309 20:26:59.191937 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:27:00 crc kubenswrapper[4750]: I0309 20:27:00.190021 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 09 20:27:00 crc kubenswrapper[4750]: I0309 20:27:00.191381 4750 scope.go:117] "RemoveContainer" containerID="724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383" Mar 09 20:27:00 crc kubenswrapper[4750]: E0309 20:27:00.192418 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:27:11 crc kubenswrapper[4750]: I0309 20:27:11.373879 4750 scope.go:117] "RemoveContainer" containerID="724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383" Mar 09 20:27:11 crc kubenswrapper[4750]: E0309 20:27:11.375377 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:27:25 crc kubenswrapper[4750]: I0309 20:27:25.373363 4750 scope.go:117] "RemoveContainer" containerID="724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383" Mar 09 20:27:25 crc kubenswrapper[4750]: E0309 20:27:25.374199 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" Mar 09 20:27:38 crc kubenswrapper[4750]: I0309 20:27:38.374617 4750 scope.go:117] "RemoveContainer" containerID="724ad6c27798b688e0628c7dad7e5126e337b120c6411fc1d44835ed1cfbe383" Mar 09 20:27:38 crc kubenswrapper[4750]: E0309 20:27:38.375901 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-scheduler\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=cinder-scheduler pod=cinder-scheduler-0_openstack(e2488a7a-65a2-442f-81bf-5c74ce640dc4)\"" pod="openstack/cinder-scheduler-0" podUID="e2488a7a-65a2-442f-81bf-5c74ce640dc4" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515153626512024453 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015153626513017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015153607651016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015153607651015466 5ustar corecore